Home
last modified time | relevance | path

Searched refs:mn (Results 1 – 35 of 35) sorted by relevance

/linux-4.4.14/mm/
Dmmu_notifier.c58 struct mmu_notifier *mn; in __mmu_notifier_release() local
66 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) in __mmu_notifier_release()
73 if (mn->ops->release) in __mmu_notifier_release()
74 mn->ops->release(mn, mm); in __mmu_notifier_release()
78 mn = hlist_entry(mm->mmu_notifier_mm->list.first, in __mmu_notifier_release()
87 hlist_del_init_rcu(&mn->hlist); in __mmu_notifier_release()
113 struct mmu_notifier *mn; in __mmu_notifier_clear_flush_young() local
117 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_flush_young()
118 if (mn->ops->clear_flush_young) in __mmu_notifier_clear_flush_young()
119 young |= mn->ops->clear_flush_young(mn, mm, start, end); in __mmu_notifier_clear_flush_young()
[all …]
Dpage_ext.c331 struct memory_notify *mn = arg; in page_ext_callback() local
336 ret = online_page_ext(mn->start_pfn, in page_ext_callback()
337 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
340 offline_page_ext(mn->start_pfn, in page_ext_callback()
341 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
344 offline_page_ext(mn->start_pfn, in page_ext_callback()
345 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
Dksm.c2049 struct memory_notify *mn = arg; in ksm_memory_callback() local
2073 ksm_check_stable_tree(mn->start_pfn, in ksm_memory_callback()
2074 mn->start_pfn + mn->nr_pages); in ksm_memory_callback()
/linux-4.4.14/drivers/gpu/drm/i915/
Di915_gem_userptr.c38 struct i915_mmu_notifier *mn; member
50 struct mmu_notifier mn; member
57 struct i915_mmu_notifier *mn; member
122 struct i915_mmu_notifier *mn = in i915_gem_userptr_mn_invalidate_range_start() local
123 container_of(_mn, struct i915_mmu_notifier, mn); in i915_gem_userptr_mn_invalidate_range_start()
129 spin_lock(&mn->lock); in i915_gem_userptr_mn_invalidate_range_start()
130 if (mn->has_linear) { in i915_gem_userptr_mn_invalidate_range_start()
131 list_for_each_entry(mo, &mn->linear, link) { in i915_gem_userptr_mn_invalidate_range_start()
140 it = interval_tree_iter_first(&mn->objects, start, end); in i915_gem_userptr_mn_invalidate_range_start()
147 spin_unlock(&mn->lock); in i915_gem_userptr_mn_invalidate_range_start()
[all …]
/linux-4.4.14/drivers/clk/qcom/
Dclk-rcg.c114 static u32 md_to_m(struct mn *mn, u32 md) in md_to_m() argument
116 md >>= mn->m_val_shift; in md_to_m()
117 md &= BIT(mn->width) - 1; in md_to_m()
140 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) in mn_to_md() argument
144 mask_w = BIT(mn->width) - 1; in mn_to_md()
145 mask = (mask_w << mn->m_val_shift) | mask_w; in mn_to_md()
149 m <<= mn->m_val_shift; in mn_to_md()
157 static u32 ns_m_to_n(struct mn *mn, u32 ns, u32 m) in ns_m_to_n() argument
159 ns = ~ns >> mn->n_val_shift; in ns_m_to_n()
160 ns &= BIT(mn->width) - 1; in ns_m_to_n()
[all …]
Dclk-rcg.h48 struct mn { struct
98 struct mn mn; member
137 struct mn mn[2]; member
Dgcc-msm8660.c111 .mn = {
162 .mn = {
213 .mn = {
264 .mn = {
315 .mn = {
366 .mn = {
417 .mn = {
468 .mn = {
517 .mn = {
566 .mn = {
[all …]
Dgcc-ipq806x.c272 .mn = {
323 .mn = {
374 .mn = {
425 .mn = {
476 .mn = {
527 .mn = {
591 .mn = {
640 .mn = {
689 .mn = {
738 .mn = {
[all …]
Dgcc-msm8960.c178 .mn = {
229 .mn = {
280 .mn = {
331 .mn = {
382 .mn = {
433 .mn = {
484 .mn = {
535 .mn = {
584 .mn = {
633 .mn = {
[all …]
Dmmcc-msm8960.c180 .mn = {
229 .mn = {
278 .mn = {
334 .mn = {
398 .mn = {
462 .mn = {
713 .mn = {
815 .mn[0] = {
823 .mn[1] = {
875 .mn[0] = {
[all …]
Dlcc-ipq806x.c119 .mn = {
233 .mn = {
313 .mn = {
372 .mn = {
Dlcc-msm8960.c100 .mn = {
206 .mn = { \
350 .mn = {
418 .mn = {
/linux-4.4.14/drivers/misc/sgi-gru/
Dgrutlbpurge.c222 static void gru_invalidate_range_start(struct mmu_notifier *mn, in gru_invalidate_range_start() argument
226 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_start()
236 static void gru_invalidate_range_end(struct mmu_notifier *mn, in gru_invalidate_range_end() argument
240 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_end()
250 static void gru_invalidate_page(struct mmu_notifier *mn, struct mm_struct *mm, in gru_invalidate_page() argument
253 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_page()
261 static void gru_release(struct mmu_notifier *mn, struct mm_struct *mm) in gru_release() argument
263 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_release()
282 struct mmu_notifier *mn, *gru_mn = NULL; in mmu_find_ops() local
286 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, in mmu_find_ops()
[all …]
/linux-4.4.14/drivers/gpu/drm/radeon/
Dradeon_mn.c43 struct mmu_notifier mn; member
83 bo->mn = NULL; in radeon_mn_destroy()
90 mmu_notifier_unregister(&rmn->mn, rmn->mm); in radeon_mn_destroy()
102 static void radeon_mn_release(struct mmu_notifier *mn, in radeon_mn_release() argument
105 struct radeon_mn *rmn = container_of(mn, struct radeon_mn, mn); in radeon_mn_release()
121 static void radeon_mn_invalidate_range_start(struct mmu_notifier *mn, in radeon_mn_invalidate_range_start() argument
126 struct radeon_mn *rmn = container_of(mn, struct radeon_mn, mn); in radeon_mn_invalidate_range_start()
204 rmn->mn.ops = &radeon_mn_ops; in radeon_mn_get()
208 r = __mmu_notifier_register(&rmn->mn, mm); in radeon_mn_get()
271 bo->mn = rmn; in radeon_mn_register()
[all …]
Dradeon.h509 struct radeon_mn *mn; member
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_mn.c43 struct mmu_notifier mn; member
83 bo->mn = NULL; in amdgpu_mn_destroy()
90 mmu_notifier_unregister(&rmn->mn, rmn->mm); in amdgpu_mn_destroy()
102 static void amdgpu_mn_release(struct mmu_notifier *mn, in amdgpu_mn_release() argument
105 struct amdgpu_mn *rmn = container_of(mn, struct amdgpu_mn, mn); in amdgpu_mn_release()
121 static void amdgpu_mn_invalidate_range_start(struct mmu_notifier *mn, in amdgpu_mn_invalidate_range_start() argument
126 struct amdgpu_mn *rmn = container_of(mn, struct amdgpu_mn, mn); in amdgpu_mn_invalidate_range_start()
205 rmn->mn.ops = &amdgpu_mn_ops; in amdgpu_mn_get()
209 r = __mmu_notifier_register(&rmn->mn, mm); in amdgpu_mn_get()
272 bo->mn = rmn; in amdgpu_mn_register()
[all …]
Damdgpu.h546 struct amdgpu_mn *mn; member
/linux-4.4.14/include/linux/
Dmmu_notifier.h51 void (*release)(struct mmu_notifier *mn,
63 int (*clear_flush_young)(struct mmu_notifier *mn,
73 int (*clear_young)(struct mmu_notifier *mn,
84 int (*test_young)(struct mmu_notifier *mn,
92 void (*change_pte)(struct mmu_notifier *mn,
104 void (*invalidate_page)(struct mmu_notifier *mn,
151 void (*invalidate_range_start)(struct mmu_notifier *mn,
154 void (*invalidate_range_end)(struct mmu_notifier *mn,
178 void (*invalidate_range)(struct mmu_notifier *mn, struct mm_struct *mm,
203 extern int mmu_notifier_register(struct mmu_notifier *mn,
[all …]
Dnvme.h103 char mn[40]; member
/linux-4.4.14/drivers/infiniband/core/
Dumem_odp.c140 static void ib_umem_notifier_release(struct mmu_notifier *mn, in ib_umem_notifier_release() argument
143 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_release()
166 static void ib_umem_notifier_invalidate_page(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_page() argument
170 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_page()
192 static void ib_umem_notifier_invalidate_range_start(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_range_start() argument
197 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_range_start()
217 static void ib_umem_notifier_invalidate_range_end(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_range_end() argument
222 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_range_end()
311 INIT_HLIST_NODE(&context->mn.hlist); in ib_umem_odp_get()
312 context->mn.ops = &ib_umem_notifiers; in ib_umem_odp_get()
[all …]
/linux-4.4.14/drivers/iommu/
Damd_iommu_v2.c51 struct mmu_notifier mn; /* mmu_notifier handle */ member
346 mmu_notifier_unregister(&pasid_state->mn, pasid_state->mm); in free_pasid_states()
365 static struct pasid_state *mn_to_state(struct mmu_notifier *mn) in mn_to_state() argument
367 return container_of(mn, struct pasid_state, mn); in mn_to_state()
370 static void __mn_flush_page(struct mmu_notifier *mn, in __mn_flush_page() argument
376 pasid_state = mn_to_state(mn); in __mn_flush_page()
382 static int mn_clear_flush_young(struct mmu_notifier *mn, in mn_clear_flush_young() argument
388 __mn_flush_page(mn, start); in mn_clear_flush_young()
393 static void mn_invalidate_page(struct mmu_notifier *mn, in mn_invalidate_page() argument
397 __mn_flush_page(mn, address); in mn_invalidate_page()
[all …]
Dintel-svm.c211 static void intel_change_pte(struct mmu_notifier *mn, struct mm_struct *mm, in intel_change_pte() argument
214 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_change_pte()
219 static void intel_invalidate_page(struct mmu_notifier *mn, struct mm_struct *mm, in intel_invalidate_page() argument
222 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_invalidate_page()
228 static void intel_invalidate_range(struct mmu_notifier *mn, in intel_invalidate_range() argument
232 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_invalidate_range()
249 static void intel_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) in intel_mm_release() argument
251 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_mm_release()
/linux-4.4.14/drivers/scsi/qla2xxx/
Dqla_bsg.c976 struct verify_chip_entry_84xx *mn = NULL; in qla84xx_updatefw() local
1022 mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_updatefw()
1023 if (!mn) { in qla84xx_updatefw()
1033 memset(mn, 0, sizeof(struct access_chip_84xx)); in qla84xx_updatefw()
1034 mn->entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_updatefw()
1035 mn->entry_count = 1; in qla84xx_updatefw()
1041 mn->options = cpu_to_le16(options); in qla84xx_updatefw()
1042 mn->fw_ver = cpu_to_le32(fw_ver); in qla84xx_updatefw()
1043 mn->fw_size = cpu_to_le32(data_len); in qla84xx_updatefw()
1044 mn->fw_seq_size = cpu_to_le32(data_len); in qla84xx_updatefw()
[all …]
Dqla_mbx.c3703 struct cs84xx_mgmt_cmd *mn; in qla84xx_verify_chip() local
3712 mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_verify_chip()
3713 if (mn == NULL) { in qla84xx_verify_chip()
3726 memset(mn, 0, sizeof(*mn)); in qla84xx_verify_chip()
3727 mn->p.req.entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_verify_chip()
3728 mn->p.req.entry_count = 1; in qla84xx_verify_chip()
3729 mn->p.req.options = cpu_to_le16(options); in qla84xx_verify_chip()
3734 (uint8_t *)mn, sizeof(*mn)); in qla84xx_verify_chip()
3736 rval = qla2x00_issue_iocb_timeout(vha, mn, mn_dma, 0, 120); in qla84xx_verify_chip()
3746 (uint8_t *)mn, sizeof(*mn)); in qla84xx_verify_chip()
[all …]
/linux-4.4.14/drivers/xen/
Dgntdev.c72 struct mmu_notifier mn; member
468 static void mn_invl_range_start(struct mmu_notifier *mn, in mn_invl_range_start() argument
472 struct gntdev_priv *priv = container_of(mn, struct gntdev_priv, mn); in mn_invl_range_start()
485 static void mn_invl_page(struct mmu_notifier *mn, in mn_invl_page() argument
489 mn_invl_range_start(mn, mm, address, address + PAGE_SIZE); in mn_invl_page()
492 static void mn_release(struct mmu_notifier *mn, in mn_release() argument
495 struct gntdev_priv *priv = container_of(mn, struct gntdev_priv, mn); in mn_release()
548 priv->mn.ops = &gntdev_mmu_ops; in gntdev_open()
549 ret = mmu_notifier_register(&priv->mn, priv->mm); in gntdev_open()
581 mmu_notifier_unregister(&priv->mn, priv->mm); in gntdev_release()
/linux-4.4.14/drivers/char/
Dmmtimer.c557 struct mmtimer_node *mn = &timers[nodeid]; in mmtimer_tasklet() local
563 spin_lock_irqsave(&mn->lock, flags); in mmtimer_tasklet()
564 if (!mn->next) in mmtimer_tasklet()
567 x = rb_entry(mn->next, struct mmtimer, list); in mmtimer_tasklet()
575 mn->next = rb_next(&x->list); in mmtimer_tasklet()
576 rb_erase(&x->list, &mn->timer_head); in mmtimer_tasklet()
595 spin_unlock_irqrestore(&mn->lock, flags); in mmtimer_tasklet()
/linux-4.4.14/net/sched/
Dsch_teql.c231 struct neighbour *mn; in __teql_resolve() local
233 mn = __neigh_lookup_errno(n->tbl, n->primary_key, dev); in __teql_resolve()
235 if (IS_ERR(mn)) in __teql_resolve()
236 return PTR_ERR(mn); in __teql_resolve()
237 n = mn; in __teql_resolve()
/linux-4.4.14/virt/kvm/
Dkvm_main.c268 static inline struct kvm *mmu_notifier_to_kvm(struct mmu_notifier *mn) in mmu_notifier_to_kvm() argument
270 return container_of(mn, struct kvm, mmu_notifier); in mmu_notifier_to_kvm()
273 static void kvm_mmu_notifier_invalidate_page(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_page() argument
277 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_page()
314 static void kvm_mmu_notifier_change_pte(struct mmu_notifier *mn, in kvm_mmu_notifier_change_pte() argument
319 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_change_pte()
330 static void kvm_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_range_start() argument
335 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_start()
356 static void kvm_mmu_notifier_invalidate_range_end(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_range_end() argument
361 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_end()
[all …]
/linux-4.4.14/drivers/gpu/drm/amd/amdkfd/
Dkfd_process.c221 static void kfd_process_notifier_release(struct mmu_notifier *mn, in kfd_process_notifier_release() argument
231 p = container_of(mn, struct kfd_process, mmu_notifier); in kfd_process_notifier_release()
/linux-4.4.14/drivers/video/fbdev/
Dpxa168fb.h220 #define CFG_HWC_1BITENA(mn) ((mn) << 25) argument
/linux-4.4.14/drivers/misc/mic/scif/
Dscif_dma.c200 static void scif_mmu_notifier_release(struct mmu_notifier *mn, in scif_mmu_notifier_release() argument
205 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_release()
210 static void scif_mmu_notifier_invalidate_page(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_page() argument
216 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_invalidate_page()
220 static void scif_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_range_start() argument
227 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_invalidate_range_start()
231 static void scif_mmu_notifier_invalidate_range_end(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_range_end() argument
/linux-4.4.14/drivers/video/fbdev/mmp/hw/
Dmmp_ctrl.h444 #define CFG_HWC_1BITENA(mn) ((mn)<<25) argument
/linux-4.4.14/arch/ia64/include/asm/
Dpal.h395 mn : 1, /* Min. state save member
707 #define pmci_proc_min_state_save_area_regd pme_processor.mn
/linux-4.4.14/drivers/nvme/host/
Dpci.c2637 memcpy(dev->model, ctrl->mn, sizeof(ctrl->mn)); in nvme_dev_add()
/linux-4.4.14/include/rdma/
Dib_verbs.h1258 struct mmu_notifier mn; member