/linux-4.4.14/mm/ |
D | mmu_notifier.c | 58 struct mmu_notifier *mn; in __mmu_notifier_release() local 66 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) in __mmu_notifier_release() 73 if (mn->ops->release) in __mmu_notifier_release() 74 mn->ops->release(mn, mm); in __mmu_notifier_release() 78 mn = hlist_entry(mm->mmu_notifier_mm->list.first, in __mmu_notifier_release() 87 hlist_del_init_rcu(&mn->hlist); in __mmu_notifier_release() 113 struct mmu_notifier *mn; in __mmu_notifier_clear_flush_young() local 117 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_flush_young() 118 if (mn->ops->clear_flush_young) in __mmu_notifier_clear_flush_young() 119 young |= mn->ops->clear_flush_young(mn, mm, start, end); in __mmu_notifier_clear_flush_young() [all …]
|
D | page_ext.c | 331 struct memory_notify *mn = arg; in page_ext_callback() local 336 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 337 mn->nr_pages, mn->status_change_nid); in page_ext_callback() 340 offline_page_ext(mn->start_pfn, in page_ext_callback() 341 mn->nr_pages, mn->status_change_nid); in page_ext_callback() 344 offline_page_ext(mn->start_pfn, in page_ext_callback() 345 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
|
D | ksm.c | 2049 struct memory_notify *mn = arg; in ksm_memory_callback() local 2073 ksm_check_stable_tree(mn->start_pfn, in ksm_memory_callback() 2074 mn->start_pfn + mn->nr_pages); in ksm_memory_callback()
|
/linux-4.4.14/drivers/gpu/drm/i915/ |
D | i915_gem_userptr.c | 38 struct i915_mmu_notifier *mn; member 50 struct mmu_notifier mn; member 57 struct i915_mmu_notifier *mn; member 122 struct i915_mmu_notifier *mn = in i915_gem_userptr_mn_invalidate_range_start() local 123 container_of(_mn, struct i915_mmu_notifier, mn); in i915_gem_userptr_mn_invalidate_range_start() 129 spin_lock(&mn->lock); in i915_gem_userptr_mn_invalidate_range_start() 130 if (mn->has_linear) { in i915_gem_userptr_mn_invalidate_range_start() 131 list_for_each_entry(mo, &mn->linear, link) { in i915_gem_userptr_mn_invalidate_range_start() 140 it = interval_tree_iter_first(&mn->objects, start, end); in i915_gem_userptr_mn_invalidate_range_start() 147 spin_unlock(&mn->lock); in i915_gem_userptr_mn_invalidate_range_start() [all …]
|
/linux-4.4.14/drivers/clk/qcom/ |
D | clk-rcg.c | 114 static u32 md_to_m(struct mn *mn, u32 md) in md_to_m() argument 116 md >>= mn->m_val_shift; in md_to_m() 117 md &= BIT(mn->width) - 1; in md_to_m() 140 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) in mn_to_md() argument 144 mask_w = BIT(mn->width) - 1; in mn_to_md() 145 mask = (mask_w << mn->m_val_shift) | mask_w; in mn_to_md() 149 m <<= mn->m_val_shift; in mn_to_md() 157 static u32 ns_m_to_n(struct mn *mn, u32 ns, u32 m) in ns_m_to_n() argument 159 ns = ~ns >> mn->n_val_shift; in ns_m_to_n() 160 ns &= BIT(mn->width) - 1; in ns_m_to_n() [all …]
|
D | clk-rcg.h | 48 struct mn { struct 98 struct mn mn; member 137 struct mn mn[2]; member
|
D | gcc-msm8660.c | 111 .mn = { 162 .mn = { 213 .mn = { 264 .mn = { 315 .mn = { 366 .mn = { 417 .mn = { 468 .mn = { 517 .mn = { 566 .mn = { [all …]
|
D | gcc-ipq806x.c | 272 .mn = { 323 .mn = { 374 .mn = { 425 .mn = { 476 .mn = { 527 .mn = { 591 .mn = { 640 .mn = { 689 .mn = { 738 .mn = { [all …]
|
D | gcc-msm8960.c | 178 .mn = { 229 .mn = { 280 .mn = { 331 .mn = { 382 .mn = { 433 .mn = { 484 .mn = { 535 .mn = { 584 .mn = { 633 .mn = { [all …]
|
D | mmcc-msm8960.c | 180 .mn = { 229 .mn = { 278 .mn = { 334 .mn = { 398 .mn = { 462 .mn = { 713 .mn = { 815 .mn[0] = { 823 .mn[1] = { 875 .mn[0] = { [all …]
|
D | lcc-ipq806x.c | 119 .mn = { 233 .mn = { 313 .mn = { 372 .mn = {
|
D | lcc-msm8960.c | 100 .mn = { 206 .mn = { \ 350 .mn = { 418 .mn = {
|
/linux-4.4.14/drivers/misc/sgi-gru/ |
D | grutlbpurge.c | 222 static void gru_invalidate_range_start(struct mmu_notifier *mn, in gru_invalidate_range_start() argument 226 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_start() 236 static void gru_invalidate_range_end(struct mmu_notifier *mn, in gru_invalidate_range_end() argument 240 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_end() 250 static void gru_invalidate_page(struct mmu_notifier *mn, struct mm_struct *mm, in gru_invalidate_page() argument 253 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_page() 261 static void gru_release(struct mmu_notifier *mn, struct mm_struct *mm) in gru_release() argument 263 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_release() 282 struct mmu_notifier *mn, *gru_mn = NULL; in mmu_find_ops() local 286 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, in mmu_find_ops() [all …]
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
D | radeon_mn.c | 43 struct mmu_notifier mn; member 83 bo->mn = NULL; in radeon_mn_destroy() 90 mmu_notifier_unregister(&rmn->mn, rmn->mm); in radeon_mn_destroy() 102 static void radeon_mn_release(struct mmu_notifier *mn, in radeon_mn_release() argument 105 struct radeon_mn *rmn = container_of(mn, struct radeon_mn, mn); in radeon_mn_release() 121 static void radeon_mn_invalidate_range_start(struct mmu_notifier *mn, in radeon_mn_invalidate_range_start() argument 126 struct radeon_mn *rmn = container_of(mn, struct radeon_mn, mn); in radeon_mn_invalidate_range_start() 204 rmn->mn.ops = &radeon_mn_ops; in radeon_mn_get() 208 r = __mmu_notifier_register(&rmn->mn, mm); in radeon_mn_get() 271 bo->mn = rmn; in radeon_mn_register() [all …]
|
D | radeon.h | 509 struct radeon_mn *mn; member
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_mn.c | 43 struct mmu_notifier mn; member 83 bo->mn = NULL; in amdgpu_mn_destroy() 90 mmu_notifier_unregister(&rmn->mn, rmn->mm); in amdgpu_mn_destroy() 102 static void amdgpu_mn_release(struct mmu_notifier *mn, in amdgpu_mn_release() argument 105 struct amdgpu_mn *rmn = container_of(mn, struct amdgpu_mn, mn); in amdgpu_mn_release() 121 static void amdgpu_mn_invalidate_range_start(struct mmu_notifier *mn, in amdgpu_mn_invalidate_range_start() argument 126 struct amdgpu_mn *rmn = container_of(mn, struct amdgpu_mn, mn); in amdgpu_mn_invalidate_range_start() 205 rmn->mn.ops = &amdgpu_mn_ops; in amdgpu_mn_get() 209 r = __mmu_notifier_register(&rmn->mn, mm); in amdgpu_mn_get() 272 bo->mn = rmn; in amdgpu_mn_register() [all …]
|
D | amdgpu.h | 546 struct amdgpu_mn *mn; member
|
/linux-4.4.14/include/linux/ |
D | mmu_notifier.h | 51 void (*release)(struct mmu_notifier *mn, 63 int (*clear_flush_young)(struct mmu_notifier *mn, 73 int (*clear_young)(struct mmu_notifier *mn, 84 int (*test_young)(struct mmu_notifier *mn, 92 void (*change_pte)(struct mmu_notifier *mn, 104 void (*invalidate_page)(struct mmu_notifier *mn, 151 void (*invalidate_range_start)(struct mmu_notifier *mn, 154 void (*invalidate_range_end)(struct mmu_notifier *mn, 178 void (*invalidate_range)(struct mmu_notifier *mn, struct mm_struct *mm, 203 extern int mmu_notifier_register(struct mmu_notifier *mn, [all …]
|
D | nvme.h | 103 char mn[40]; member
|
/linux-4.4.14/drivers/infiniband/core/ |
D | umem_odp.c | 140 static void ib_umem_notifier_release(struct mmu_notifier *mn, in ib_umem_notifier_release() argument 143 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_release() 166 static void ib_umem_notifier_invalidate_page(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_page() argument 170 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_page() 192 static void ib_umem_notifier_invalidate_range_start(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_range_start() argument 197 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_range_start() 217 static void ib_umem_notifier_invalidate_range_end(struct mmu_notifier *mn, in ib_umem_notifier_invalidate_range_end() argument 222 struct ib_ucontext *context = container_of(mn, struct ib_ucontext, mn); in ib_umem_notifier_invalidate_range_end() 311 INIT_HLIST_NODE(&context->mn.hlist); in ib_umem_odp_get() 312 context->mn.ops = &ib_umem_notifiers; in ib_umem_odp_get() [all …]
|
/linux-4.4.14/drivers/iommu/ |
D | amd_iommu_v2.c | 51 struct mmu_notifier mn; /* mmu_notifier handle */ member 346 mmu_notifier_unregister(&pasid_state->mn, pasid_state->mm); in free_pasid_states() 365 static struct pasid_state *mn_to_state(struct mmu_notifier *mn) in mn_to_state() argument 367 return container_of(mn, struct pasid_state, mn); in mn_to_state() 370 static void __mn_flush_page(struct mmu_notifier *mn, in __mn_flush_page() argument 376 pasid_state = mn_to_state(mn); in __mn_flush_page() 382 static int mn_clear_flush_young(struct mmu_notifier *mn, in mn_clear_flush_young() argument 388 __mn_flush_page(mn, start); in mn_clear_flush_young() 393 static void mn_invalidate_page(struct mmu_notifier *mn, in mn_invalidate_page() argument 397 __mn_flush_page(mn, address); in mn_invalidate_page() [all …]
|
D | intel-svm.c | 211 static void intel_change_pte(struct mmu_notifier *mn, struct mm_struct *mm, in intel_change_pte() argument 214 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_change_pte() 219 static void intel_invalidate_page(struct mmu_notifier *mn, struct mm_struct *mm, in intel_invalidate_page() argument 222 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_invalidate_page() 228 static void intel_invalidate_range(struct mmu_notifier *mn, in intel_invalidate_range() argument 232 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_invalidate_range() 249 static void intel_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) in intel_mm_release() argument 251 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); in intel_mm_release()
|
/linux-4.4.14/drivers/scsi/qla2xxx/ |
D | qla_bsg.c | 976 struct verify_chip_entry_84xx *mn = NULL; in qla84xx_updatefw() local 1022 mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_updatefw() 1023 if (!mn) { in qla84xx_updatefw() 1033 memset(mn, 0, sizeof(struct access_chip_84xx)); in qla84xx_updatefw() 1034 mn->entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_updatefw() 1035 mn->entry_count = 1; in qla84xx_updatefw() 1041 mn->options = cpu_to_le16(options); in qla84xx_updatefw() 1042 mn->fw_ver = cpu_to_le32(fw_ver); in qla84xx_updatefw() 1043 mn->fw_size = cpu_to_le32(data_len); in qla84xx_updatefw() 1044 mn->fw_seq_size = cpu_to_le32(data_len); in qla84xx_updatefw() [all …]
|
D | qla_mbx.c | 3703 struct cs84xx_mgmt_cmd *mn; in qla84xx_verify_chip() local 3712 mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_verify_chip() 3713 if (mn == NULL) { in qla84xx_verify_chip() 3726 memset(mn, 0, sizeof(*mn)); in qla84xx_verify_chip() 3727 mn->p.req.entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_verify_chip() 3728 mn->p.req.entry_count = 1; in qla84xx_verify_chip() 3729 mn->p.req.options = cpu_to_le16(options); in qla84xx_verify_chip() 3734 (uint8_t *)mn, sizeof(*mn)); in qla84xx_verify_chip() 3736 rval = qla2x00_issue_iocb_timeout(vha, mn, mn_dma, 0, 120); in qla84xx_verify_chip() 3746 (uint8_t *)mn, sizeof(*mn)); in qla84xx_verify_chip() [all …]
|
/linux-4.4.14/drivers/xen/ |
D | gntdev.c | 72 struct mmu_notifier mn; member 468 static void mn_invl_range_start(struct mmu_notifier *mn, in mn_invl_range_start() argument 472 struct gntdev_priv *priv = container_of(mn, struct gntdev_priv, mn); in mn_invl_range_start() 485 static void mn_invl_page(struct mmu_notifier *mn, in mn_invl_page() argument 489 mn_invl_range_start(mn, mm, address, address + PAGE_SIZE); in mn_invl_page() 492 static void mn_release(struct mmu_notifier *mn, in mn_release() argument 495 struct gntdev_priv *priv = container_of(mn, struct gntdev_priv, mn); in mn_release() 548 priv->mn.ops = &gntdev_mmu_ops; in gntdev_open() 549 ret = mmu_notifier_register(&priv->mn, priv->mm); in gntdev_open() 581 mmu_notifier_unregister(&priv->mn, priv->mm); in gntdev_release()
|
/linux-4.4.14/drivers/char/ |
D | mmtimer.c | 557 struct mmtimer_node *mn = &timers[nodeid]; in mmtimer_tasklet() local 563 spin_lock_irqsave(&mn->lock, flags); in mmtimer_tasklet() 564 if (!mn->next) in mmtimer_tasklet() 567 x = rb_entry(mn->next, struct mmtimer, list); in mmtimer_tasklet() 575 mn->next = rb_next(&x->list); in mmtimer_tasklet() 576 rb_erase(&x->list, &mn->timer_head); in mmtimer_tasklet() 595 spin_unlock_irqrestore(&mn->lock, flags); in mmtimer_tasklet()
|
/linux-4.4.14/net/sched/ |
D | sch_teql.c | 231 struct neighbour *mn; in __teql_resolve() local 233 mn = __neigh_lookup_errno(n->tbl, n->primary_key, dev); in __teql_resolve() 235 if (IS_ERR(mn)) in __teql_resolve() 236 return PTR_ERR(mn); in __teql_resolve() 237 n = mn; in __teql_resolve()
|
/linux-4.4.14/virt/kvm/ |
D | kvm_main.c | 268 static inline struct kvm *mmu_notifier_to_kvm(struct mmu_notifier *mn) in mmu_notifier_to_kvm() argument 270 return container_of(mn, struct kvm, mmu_notifier); in mmu_notifier_to_kvm() 273 static void kvm_mmu_notifier_invalidate_page(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_page() argument 277 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_page() 314 static void kvm_mmu_notifier_change_pte(struct mmu_notifier *mn, in kvm_mmu_notifier_change_pte() argument 319 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_change_pte() 330 static void kvm_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_range_start() argument 335 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_start() 356 static void kvm_mmu_notifier_invalidate_range_end(struct mmu_notifier *mn, in kvm_mmu_notifier_invalidate_range_end() argument 361 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_end() [all …]
|
/linux-4.4.14/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_process.c | 221 static void kfd_process_notifier_release(struct mmu_notifier *mn, in kfd_process_notifier_release() argument 231 p = container_of(mn, struct kfd_process, mmu_notifier); in kfd_process_notifier_release()
|
/linux-4.4.14/drivers/video/fbdev/ |
D | pxa168fb.h | 220 #define CFG_HWC_1BITENA(mn) ((mn) << 25) argument
|
/linux-4.4.14/drivers/misc/mic/scif/ |
D | scif_dma.c | 200 static void scif_mmu_notifier_release(struct mmu_notifier *mn, in scif_mmu_notifier_release() argument 205 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_release() 210 static void scif_mmu_notifier_invalidate_page(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_page() argument 216 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_invalidate_page() 220 static void scif_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_range_start() argument 227 mmn = container_of(mn, struct scif_mmu_notif, ep_mmu_notifier); in scif_mmu_notifier_invalidate_range_start() 231 static void scif_mmu_notifier_invalidate_range_end(struct mmu_notifier *mn, in scif_mmu_notifier_invalidate_range_end() argument
|
/linux-4.4.14/drivers/video/fbdev/mmp/hw/ |
D | mmp_ctrl.h | 444 #define CFG_HWC_1BITENA(mn) ((mn)<<25) argument
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | pal.h | 395 mn : 1, /* Min. state save member 707 #define pmci_proc_min_state_save_area_regd pme_processor.mn
|
/linux-4.4.14/drivers/nvme/host/ |
D | pci.c | 2637 memcpy(dev->model, ctrl->mn, sizeof(ctrl->mn)); in nvme_dev_add()
|
/linux-4.4.14/include/rdma/ |
D | ib_verbs.h | 1258 struct mmu_notifier mn; member
|