dmb 49 arch/arm/common/mcpm_entry.c dmb(); dmb 65 arch/arm/common/mcpm_entry.c dmb(); dmb 301 arch/arm/include/asm/assembler.h ALT_SMP(dmb ish) dmb 303 arch/arm/include/asm/assembler.h ALT_SMP(W(dmb) ish) dmb 306 arch/arm/include/asm/assembler.h ALT_SMP(mcr p15, 0, r0, c7, c10, 5) @ dmb dmb 67 arch/arm/include/asm/barrier.h #define dma_rmb() dmb(osh) dmb 68 arch/arm/include/asm/barrier.h #define dma_wmb() dmb(oshst) dmb 77 arch/arm/include/asm/barrier.h #define __smp_mb() dmb(ish) dmb 79 arch/arm/include/asm/barrier.h #define __smp_wmb() dmb(ishst) dmb 153 arch/arm/kernel/smp_tlb.c dmb(); dmb 75 arch/arm/mach-bcm/platsmp-brcmstb.c dmb(); dmb 66 arch/arm/mm/cache-b15-rac.c dmb(); dmb 80 arch/arm/mm/cache-b15-rac.c dmb(); dmb 92 arch/arm64/include/asm/assembler.h dmb \opt dmb 102 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC_OP_RETURN( , dmb ish, , l, "memory", __VA_ARGS__)\ dmb 106 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\ dmb 117 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\ dmb 201 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC64_OP_RETURN(, dmb ish, , l, "memory", __VA_ARGS__) \ dmb 205 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \ dmb 216 arch/arm64/include/asm/atomic_ll_sc.h ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \ dmb 311 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, b, mb_, 8, dmb ish, , l, "memory", K) dmb 312 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, h, mb_, 16, dmb ish, , l, "memory", K) dmb 313 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, , mb_, 32, dmb ish, , l, "memory", K) dmb 314 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE( , , mb_, 64, dmb ish, , l, "memory", L) dmb 348 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_DBL(_mb, dmb ish, l, "memory") dmb 36 arch/arm64/include/asm/barrier.h #define dma_rmb() dmb(oshld) dmb 37 arch/arm64/include/asm/barrier.h #define dma_wmb() dmb(oshst) dmb 60 arch/arm64/include/asm/barrier.h #define __smp_mb() dmb(ish) dmb 61 arch/arm64/include/asm/barrier.h #define __smp_rmb() dmb(ishld) dmb 62 arch/arm64/include/asm/barrier.h #define __smp_wmb() dmb(ishst) dmb 57 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, b, mb_, 8, dmb ish, nop, , a, l, "memory") dmb 58 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, h, mb_, 16, dmb ish, nop, , a, l, "memory") dmb 59 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, , mb_, 32, dmb ish, nop, , a, l, "memory") dmb 60 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE( , , mb_, 64, dmb ish, nop, , a, l, "memory") dmb 17 arch/arm64/include/asm/vdso/compat_barrier.h #ifdef dmb dmb 24 arch/arm64/include/asm/vdso/compat_barrier.h #define aarch32_smp_mb() dmb(ish) dmb 25 arch/arm64/include/asm/vdso/compat_barrier.h #define aarch32_smp_rmb() dmb(ishld) dmb 26 arch/arm64/include/asm/vdso/compat_barrier.h #define aarch32_smp_wmb() dmb(ishst) dmb 28 arch/arm64/include/asm/vdso/compat_barrier.h #define aarch32_smp_mb() dmb(ish) dmb 30 arch/arm64/include/asm/vdso/compat_barrier.h #define aarch32_smp_wmb() dmb(ishst) dmb 484 arch/arm64/kernel/armv8_deprecated.c dmb(sy); dmb 84 arch/arm64/mm/flush.c dmb(osh); dmb 929 drivers/clk/tegra/clk-tegra20.c dmb(); dmb 1103 drivers/clk/tegra/clk-tegra30.c dmb(); dmb 828 drivers/irqchip/irq-gic.c dmb(ishst); dmb 298 drivers/irqchip/irq-hip04.c dmb(ishst); dmb 117 drivers/s390/net/ism.h u64 dmb; dmb 208 drivers/s390/net/ism.h #define ISM_CREATE_REQ(dmb, idx, sf, offset) \ dmb 209 drivers/s390/net/ism.h ((dmb) | (idx) << 24 | (sf) << 23 | (offset)) dmb 216 drivers/s390/net/ism_drv.c static void ism_free_dmb(struct ism_dev *ism, struct smcd_dmb *dmb) dmb 218 drivers/s390/net/ism_drv.c clear_bit(dmb->sba_idx, ism->sba_bitmap); dmb 219 drivers/s390/net/ism_drv.c dma_free_coherent(&ism->pdev->dev, dmb->dmb_len, dmb 220 drivers/s390/net/ism_drv.c dmb->cpu_addr, dmb->dma_addr); dmb 223 drivers/s390/net/ism_drv.c static int ism_alloc_dmb(struct ism_dev *ism, struct smcd_dmb *dmb) dmb 227 drivers/s390/net/ism_drv.c if (PAGE_ALIGN(dmb->dmb_len) > dma_get_max_seg_size(&ism->pdev->dev)) dmb 230 drivers/s390/net/ism_drv.c if (!dmb->sba_idx) { dmb 236 drivers/s390/net/ism_drv.c dmb->sba_idx = bit; dmb 238 drivers/s390/net/ism_drv.c if (dmb->sba_idx < ISM_DMB_BIT_OFFSET || dmb 239 drivers/s390/net/ism_drv.c test_and_set_bit(dmb->sba_idx, ism->sba_bitmap)) dmb 242 drivers/s390/net/ism_drv.c dmb->cpu_addr = dma_alloc_coherent(&ism->pdev->dev, dmb->dmb_len, dmb 243 drivers/s390/net/ism_drv.c &dmb->dma_addr, dmb 245 drivers/s390/net/ism_drv.c if (!dmb->cpu_addr) dmb 246 drivers/s390/net/ism_drv.c clear_bit(dmb->sba_idx, ism->sba_bitmap); dmb 248 drivers/s390/net/ism_drv.c return dmb->cpu_addr ? 0 : -ENOMEM; dmb 251 drivers/s390/net/ism_drv.c static int ism_register_dmb(struct smcd_dev *smcd, struct smcd_dmb *dmb) dmb 257 drivers/s390/net/ism_drv.c ret = ism_alloc_dmb(ism, dmb); dmb 265 drivers/s390/net/ism_drv.c cmd.request.dmb = dmb->dma_addr; dmb 266 drivers/s390/net/ism_drv.c cmd.request.dmb_len = dmb->dmb_len; dmb 267 drivers/s390/net/ism_drv.c cmd.request.sba_idx = dmb->sba_idx; dmb 268 drivers/s390/net/ism_drv.c cmd.request.vlan_valid = dmb->vlan_valid; dmb 269 drivers/s390/net/ism_drv.c cmd.request.vlan_id = dmb->vlan_id; dmb 270 drivers/s390/net/ism_drv.c cmd.request.rgid = dmb->rgid; dmb 274 drivers/s390/net/ism_drv.c ism_free_dmb(ism, dmb); dmb 277 drivers/s390/net/ism_drv.c dmb->dmb_tok = cmd.response.dmb_tok; dmb 282 drivers/s390/net/ism_drv.c static int ism_unregister_dmb(struct smcd_dev *smcd, struct smcd_dmb *dmb) dmb 292 drivers/s390/net/ism_drv.c cmd.request.dmb_tok = dmb->dmb_tok; dmb 298 drivers/s390/net/ism_drv.c ism_free_dmb(ism, dmb); dmb 53 include/net/smc.h int (*register_dmb)(struct smcd_dev *dev, struct smcd_dmb *dmb); dmb 54 include/net/smc.h int (*unregister_dmb)(struct smcd_dev *dev, struct smcd_dmb *dmb); dmb 148 net/smc/smc_ism.c struct smcd_dmb dmb; dmb 150 net/smc/smc_ism.c memset(&dmb, 0, sizeof(dmb)); dmb 151 net/smc/smc_ism.c dmb.dmb_tok = dmb_desc->token; dmb 152 net/smc/smc_ism.c dmb.sba_idx = dmb_desc->sba_idx; dmb 153 net/smc/smc_ism.c dmb.cpu_addr = dmb_desc->cpu_addr; dmb 154 net/smc/smc_ism.c dmb.dma_addr = dmb_desc->dma_addr; dmb 155 net/smc/smc_ism.c dmb.dmb_len = dmb_desc->len; dmb 156 net/smc/smc_ism.c return smcd->ops->unregister_dmb(smcd, &dmb); dmb 162 net/smc/smc_ism.c struct smcd_dmb dmb; dmb 165 net/smc/smc_ism.c memset(&dmb, 0, sizeof(dmb)); dmb 166 net/smc/smc_ism.c dmb.dmb_len = dmb_len; dmb 167 net/smc/smc_ism.c dmb.sba_idx = dmb_desc->sba_idx; dmb 168 net/smc/smc_ism.c dmb.vlan_id = lgr->vlan_id; dmb 169 net/smc/smc_ism.c dmb.rgid = lgr->peer_gid; dmb 170 net/smc/smc_ism.c rc = lgr->smcd->ops->register_dmb(lgr->smcd, &dmb); dmb 172 net/smc/smc_ism.c dmb_desc->sba_idx = dmb.sba_idx; dmb 173 net/smc/smc_ism.c dmb_desc->token = dmb.dmb_tok; dmb 174 net/smc/smc_ism.c dmb_desc->cpu_addr = dmb.cpu_addr; dmb 175 net/smc/smc_ism.c dmb_desc->dma_addr = dmb.dma_addr; dmb 176 net/smc/smc_ism.c dmb_desc->len = dmb.dmb_len;