ms 197 arch/arm/mach-pxa/include/mach/regs-lcd.h #define SMART_DELAY(ms) (SMART_CMD_DELAY | ((ms) & 0xff)) ms 896 arch/ia64/kernel/mca.c const pal_min_state_area_t *ms = sos->pal_min_state; ms 903 arch/ia64/kernel/mca.c regs->cr_iip = ms->pmsa_iip; ms 904 arch/ia64/kernel/mca.c regs->cr_ipsr = ms->pmsa_ipsr; ms 905 arch/ia64/kernel/mca.c regs->cr_ifs = ms->pmsa_ifs; ms 907 arch/ia64/kernel/mca.c regs->cr_iip = ms->pmsa_xip; ms 908 arch/ia64/kernel/mca.c regs->cr_ipsr = ms->pmsa_xpsr; ms 909 arch/ia64/kernel/mca.c regs->cr_ifs = ms->pmsa_xfs; ms 911 arch/ia64/kernel/mca.c sos->iip = ms->pmsa_iip; ms 912 arch/ia64/kernel/mca.c sos->ipsr = ms->pmsa_ipsr; ms 913 arch/ia64/kernel/mca.c sos->ifs = ms->pmsa_ifs; ms 915 arch/ia64/kernel/mca.c regs->pr = ms->pmsa_pr; ms 916 arch/ia64/kernel/mca.c regs->b0 = ms->pmsa_br0; ms 917 arch/ia64/kernel/mca.c regs->ar_rsc = ms->pmsa_rsc; ms 918 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[1-1], ms->pmsa_nat_bits, ®s->r1, nat); ms 919 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[2-1], ms->pmsa_nat_bits, ®s->r2, nat); ms 920 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[3-1], ms->pmsa_nat_bits, ®s->r3, nat); ms 921 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[8-1], ms->pmsa_nat_bits, ®s->r8, nat); ms 922 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[9-1], ms->pmsa_nat_bits, ®s->r9, nat); ms 923 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[10-1], ms->pmsa_nat_bits, ®s->r10, nat); ms 924 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[11-1], ms->pmsa_nat_bits, ®s->r11, nat); ms 925 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[12-1], ms->pmsa_nat_bits, ®s->r12, nat); ms 926 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[13-1], ms->pmsa_nat_bits, ®s->r13, nat); ms 927 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[14-1], ms->pmsa_nat_bits, ®s->r14, nat); ms 928 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[15-1], ms->pmsa_nat_bits, ®s->r15, nat); ms 930 arch/ia64/kernel/mca.c bank = ms->pmsa_bank1_gr; ms 932 arch/ia64/kernel/mca.c bank = ms->pmsa_bank0_gr; ms 933 arch/ia64/kernel/mca.c copy_reg(&bank[16-16], ms->pmsa_nat_bits, ®s->r16, nat); ms 934 arch/ia64/kernel/mca.c copy_reg(&bank[17-16], ms->pmsa_nat_bits, ®s->r17, nat); ms 935 arch/ia64/kernel/mca.c copy_reg(&bank[18-16], ms->pmsa_nat_bits, ®s->r18, nat); ms 936 arch/ia64/kernel/mca.c copy_reg(&bank[19-16], ms->pmsa_nat_bits, ®s->r19, nat); ms 937 arch/ia64/kernel/mca.c copy_reg(&bank[20-16], ms->pmsa_nat_bits, ®s->r20, nat); ms 938 arch/ia64/kernel/mca.c copy_reg(&bank[21-16], ms->pmsa_nat_bits, ®s->r21, nat); ms 939 arch/ia64/kernel/mca.c copy_reg(&bank[22-16], ms->pmsa_nat_bits, ®s->r22, nat); ms 940 arch/ia64/kernel/mca.c copy_reg(&bank[23-16], ms->pmsa_nat_bits, ®s->r23, nat); ms 941 arch/ia64/kernel/mca.c copy_reg(&bank[24-16], ms->pmsa_nat_bits, ®s->r24, nat); ms 942 arch/ia64/kernel/mca.c copy_reg(&bank[25-16], ms->pmsa_nat_bits, ®s->r25, nat); ms 943 arch/ia64/kernel/mca.c copy_reg(&bank[26-16], ms->pmsa_nat_bits, ®s->r26, nat); ms 944 arch/ia64/kernel/mca.c copy_reg(&bank[27-16], ms->pmsa_nat_bits, ®s->r27, nat); ms 945 arch/ia64/kernel/mca.c copy_reg(&bank[28-16], ms->pmsa_nat_bits, ®s->r28, nat); ms 946 arch/ia64/kernel/mca.c copy_reg(&bank[29-16], ms->pmsa_nat_bits, ®s->r29, nat); ms 947 arch/ia64/kernel/mca.c copy_reg(&bank[30-16], ms->pmsa_nat_bits, ®s->r30, nat); ms 948 arch/ia64/kernel/mca.c copy_reg(&bank[31-16], ms->pmsa_nat_bits, ®s->r31, nat); ms 972 arch/ia64/kernel/mca.c const pal_min_state_area_t *ms = sos->pal_min_state; ms 982 arch/ia64/kernel/mca.c u64 r12 = ms->pmsa_gr[12-1], r13 = ms->pmsa_gr[13-1]; ms 996 arch/ia64/kernel/mca.c regs->cr_ipsr = ms->pmsa_ipsr; ms 1053 arch/ia64/kernel/mca.c if (!mca_recover_range(ms->pmsa_iip)) { ms 1112 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[4-1], ms->pmsa_nat_bits, &old_sw->r4, &old_unat); ms 1113 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[5-1], ms->pmsa_nat_bits, &old_sw->r5, &old_unat); ms 1114 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[6-1], ms->pmsa_nat_bits, &old_sw->r6, &old_unat); ms 1115 arch/ia64/kernel/mca.c copy_reg(&ms->pmsa_gr[7-1], ms->pmsa_nat_bits, &old_sw->r7, &old_unat); ms 1117 arch/ia64/kernel/mca.c old_sw->b1 = ms->pmsa_br1; ms 94 arch/mips/kernel/spinlock_test.c struct spin_multi_state ms; ms 97 arch/mips/kernel/spinlock_test.c ms.lock = __RAW_SPIN_LOCK_UNLOCKED("multi_get"); ms 98 arch/mips/kernel/spinlock_test.c ms.loops = 1000000; ms 100 arch/mips/kernel/spinlock_test.c atomic_set(&ms.start_wait, 2); ms 101 arch/mips/kernel/spinlock_test.c atomic_set(&ms.enter_wait, 2); ms 102 arch/mips/kernel/spinlock_test.c atomic_set(&ms.exit_wait, 2); ms 103 arch/mips/kernel/spinlock_test.c t1.state = &ms; ms 104 arch/mips/kernel/spinlock_test.c t2.state = &ms; ms 492 arch/powerpc/kernel/rtas.c unsigned int ms = 0; ms 495 arch/powerpc/kernel/rtas.c ms = 1; ms 499 arch/powerpc/kernel/rtas.c for (ms = 1; order > 0; order--) ms 500 arch/powerpc/kernel/rtas.c ms *= 10; ms 503 arch/powerpc/kernel/rtas.c return ms; ms 510 arch/powerpc/kernel/rtas.c unsigned int ms; ms 513 arch/powerpc/kernel/rtas.c ms = rtas_busy_delay_time(status); ms 514 arch/powerpc/kernel/rtas.c if (ms && need_resched()) ms 515 arch/powerpc/kernel/rtas.c msleep(ms); ms 517 arch/powerpc/kernel/rtas.c return ms; ms 117 arch/powerpc/sysdev/xive/spapr.c unsigned int ms = 0; ms 120 arch/powerpc/sysdev/xive/spapr.c ms = get_longbusy_msecs(rc); ms 122 arch/powerpc/sysdev/xive/spapr.c ms = 10; /* seems appropriate for XIVE hcalls */ ms 125 arch/powerpc/sysdev/xive/spapr.c return ms; ms 130 arch/powerpc/sysdev/xive/spapr.c unsigned int ms; ms 132 arch/powerpc/sysdev/xive/spapr.c ms = plpar_busy_delay_time(rc); ms 133 arch/powerpc/sysdev/xive/spapr.c if (ms) ms 134 arch/powerpc/sysdev/xive/spapr.c mdelay(ms); ms 136 arch/powerpc/sysdev/xive/spapr.c return ms; ms 58 arch/s390/include/asm/nmi.h u64 ms : 1; /* 21 psw mask and key validity */ ms 29 arch/s390/kernel/machine_kexec_file.c struct module_signature *ms; ms 44 arch/s390/kernel/machine_kexec_file.c ms = (void *)kernel + kernel_len - sizeof(*ms); ms 45 arch/s390/kernel/machine_kexec_file.c kernel_len -= sizeof(*ms); ms 47 arch/s390/kernel/machine_kexec_file.c sig_len = be32_to_cpu(ms->sig_len); ms 52 arch/s390/kernel/machine_kexec_file.c if (ms->id_type != PKEY_ID_PKCS7) ms 55 arch/s390/kernel/machine_kexec_file.c if (ms->algo != 0 || ms 56 arch/s390/kernel/machine_kexec_file.c ms->hash != 0 || ms 57 arch/s390/kernel/machine_kexec_file.c ms->signer_len != 0 || ms 58 arch/s390/kernel/machine_kexec_file.c ms->key_id_len != 0 || ms 59 arch/s390/kernel/machine_kexec_file.c ms->__pad[0] != 0 || ms 60 arch/s390/kernel/machine_kexec_file.c ms->__pad[1] != 0 || ms 61 arch/s390/kernel/machine_kexec_file.c ms->__pad[2] != 0) { ms 292 arch/s390/kernel/nmi.c if (!mci.ms || !mci.pm || !mci.ia) ms 1002 arch/s390/kvm/kvm-s390.c struct kvm_memory_slot *ms; ms 1020 arch/s390/kvm/kvm-s390.c ms = slots->memslots + slotnr; ms 1021 arch/s390/kvm/kvm-s390.c if (!ms->dirty_bitmap) ms 1029 arch/s390/kvm/kvm-s390.c memset(kvm_second_dirty_bitmap(ms), 0xff, kvm_dirty_bitmap_bytes(ms)); ms 1030 arch/s390/kvm/kvm-s390.c ram_pages += ms->npages; ms 1973 arch/s390/kvm/kvm-s390.c struct kvm_memory_slot *ms = slots->memslots + slotidx; ms 1974 arch/s390/kvm/kvm-s390.c unsigned long ofs = cur_gfn - ms->base_gfn; ms 1976 arch/s390/kvm/kvm-s390.c if (ms->base_gfn + ms->npages <= cur_gfn) { ms 1982 arch/s390/kvm/kvm-s390.c ms = slots->memslots + slotidx; ms 1985 arch/s390/kvm/kvm-s390.c ofs = find_next_bit(kvm_second_dirty_bitmap(ms), ms->npages, ofs); ms 1986 arch/s390/kvm/kvm-s390.c while ((slotidx > 0) && (ofs >= ms->npages)) { ms 1988 arch/s390/kvm/kvm-s390.c ms = slots->memslots + slotidx; ms 1989 arch/s390/kvm/kvm-s390.c ofs = find_next_bit(kvm_second_dirty_bitmap(ms), ms->npages, 0); ms 1991 arch/s390/kvm/kvm-s390.c return ms->base_gfn + ofs; ms 1999 arch/s390/kvm/kvm-s390.c struct kvm_memory_slot *ms; ms 2002 arch/s390/kvm/kvm-s390.c ms = gfn_to_memslot(kvm, cur_gfn); ms 2005 arch/s390/kvm/kvm-s390.c if (!ms) ms 2015 arch/s390/kvm/kvm-s390.c if (test_and_clear_bit(cur_gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) ms 2033 arch/s390/kvm/kvm-s390.c if (cur_gfn - ms->base_gfn >= ms->npages) { ms 2034 arch/s390/kvm/kvm-s390.c ms = gfn_to_memslot(kvm, cur_gfn); ms 2035 arch/s390/kvm/kvm-s390.c if (!ms) ms 1174 arch/s390/kvm/priv.c struct kvm_memory_slot *ms = gfn_to_memslot(vcpu->kvm, gfn); ms 1177 arch/s390/kvm/priv.c if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) ms 485 arch/x86/events/perf_event.h #define EVENT_EXTRA_REG(e, ms, m, vm, i) { \ ms 487 arch/x86/events/perf_event.h .msr = (ms), \ ms 1126 arch/x86/kernel/cpu/amd.c u32 ms; ms 1143 arch/x86/kernel/cpu/amd.c ms = (cpu->x86_model << 4) | cpu->x86_stepping; ms 1146 arch/x86/kernel/cpu/amd.c (ms >= AMD_MODEL_RANGE_START(range)) && ms 1147 arch/x86/kernel/cpu/amd.c (ms <= AMD_MODEL_RANGE_END(range))) ms 381 arch/x86/kernel/tsc.c static unsigned long pit_calibrate_tsc(u32 latch, unsigned long ms, int loopmin) ms 443 arch/x86/kernel/tsc.c do_div(delta, ms); ms 713 arch/x86/kernel/tsc.c unsigned long flags, latch, ms; ms 743 arch/x86/kernel/tsc.c ms = CAL_MS; ms 757 arch/x86/kernel/tsc.c tsc_pit_khz = pit_calibrate_tsc(latch, ms, loopmin); ms 804 arch/x86/kernel/tsc.c ms = CAL2_MS; ms 607 drivers/acpi/osl.c void acpi_os_sleep(u64 ms) ms 609 drivers/acpi/osl.c msleep(ms); ms 100 drivers/auxdisplay/charlcd.c static void long_sleep(int ms) ms 102 drivers/auxdisplay/charlcd.c schedule_timeout_interruptible(msecs_to_jiffies(ms)); ms 86 drivers/bcma/driver_chipcommon.c u32 ms) ms 91 drivers/bcma/driver_chipcommon.c ticks = bcma_chipco_watchdog_timer_set(cc, cc->ticks_per_ms * ms); ms 339 drivers/block/amiflop.c static void ms_delay(int ms) ms 344 drivers/block/amiflop.c if (ms > 0) { ms 346 drivers/block/amiflop.c ticks = MS_TICKS*ms-1; ms 1150 drivers/block/drbd/drbd_bitmap.c unsigned int ms = jiffies_to_msecs(jiffies - now); ms 1151 drivers/block/drbd/drbd_bitmap.c if (ms > 5) { ms 1154 drivers/block/drbd/drbd_bitmap.c count, ms); ms 4358 drivers/block/drbd/drbd_receiver.c union drbd_state ms; ms 4371 drivers/block/drbd/drbd_receiver.c ms.i = ps.i; ms 4373 drivers/block/drbd/drbd_receiver.c ms.conn = c_tab[ps.conn]; ms 4374 drivers/block/drbd/drbd_receiver.c ms.peer = ps.role; ms 4375 drivers/block/drbd/drbd_receiver.c ms.role = ps.peer; ms 4376 drivers/block/drbd/drbd_receiver.c ms.pdsk = ps.disk; ms 4377 drivers/block/drbd/drbd_receiver.c ms.disk = ps.pdsk; ms 4378 drivers/block/drbd/drbd_receiver.c ms.peer_isp = (ps.aftr_isp | ps.user_isp); ms 4380 drivers/block/drbd/drbd_receiver.c return ms; ms 665 drivers/block/paride/pcd.c static int pcd_probe(struct pcd_unit *cd, int ms, char *id) ms 667 drivers/block/paride/pcd.c if (ms == -1) { ms 672 drivers/block/paride/pcd.c cd->drive = ms; ms 651 drivers/block/paride/pf.c char *ms[2] = { "master", "slave" }; ms 681 drivers/block/paride/pf.c pf->name, mf, id, ms[pf->drive], pf->lun, dt); ms 440 drivers/block/paride/pg.c char *ms[2] = { "master", "slave" }; ms 455 drivers/block/paride/pg.c printk("%s: %s %s, %s\n", dev->name, mf, id, ms[dev->drive]); ms 531 drivers/block/paride/pt.c char *ms[2] = { "master", "slave" }; ms 571 drivers/block/paride/pt.c printk("%s: %s %s, %s", tape->name, mf, id, ms[tape->drive]); ms 425 drivers/clk/rockchip/clk.h #define COMPOSITE(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\ ms 435 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 446 drivers/clk/rockchip/clk.h #define COMPOSITE_DIV_OFFSET(_id, cname, pnames, f, mo, ms, mw, \ ms 456 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 505 drivers/clk/rockchip/clk.h #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ ms 515 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 523 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE(_id, cname, pnames, f, mo, ms, mw, mf, \ ms 533 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 542 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE_DIVTBL(_id, cname, pnames, f, mo, ms, \ ms 552 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 613 drivers/clk/rockchip/clk.h #define COMPOSITE_DDRCLK(_id, cname, pnames, f, mo, ms, mw, \ ms 623 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 754 drivers/clk/rockchip/clk.h #define COMPOSITE_HALFDIV(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\ ms 764 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 775 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE_HALFDIV(_id, cname, pnames, f, mo, ms, mw, mf, \ ms 785 drivers/clk/rockchip/clk.h .mux_shift = ms, \ ms 198 drivers/clocksource/timer-tegra.c u32 ms = readl_relaxed(reg_base + RTC_MILLISECONDS); ms 201 drivers/clocksource/timer-tegra.c return (u64)s * MSEC_PER_SEC + ms; ms 87 drivers/cpufreq/pmac32-cpufreq.c static inline void local_delay(unsigned long ms) ms 90 drivers/cpufreq/pmac32-cpufreq.c mdelay(ms); ms 92 drivers/cpufreq/pmac32-cpufreq.c msleep(ms); ms 279 drivers/dma/imx-sdma.c u32 ms; ms 392 drivers/gpu/drm/arm/malidp_crtc.c struct malidp_plane_state *ms = to_malidp_plane_state(pstate); ms 412 drivers/gpu/drm/arm/malidp_crtc.c if (ms->rotmem_size > rot_mem_usable) ms 138 drivers/gpu/drm/arm/malidp_planes.c struct malidp_plane_state *ms = to_malidp_plane_state(state); ms 140 drivers/gpu/drm/arm/malidp_planes.c drm_printf(p, "\trotmem_size=%u\n", ms->rotmem_size); ms 141 drivers/gpu/drm/arm/malidp_planes.c drm_printf(p, "\tformat_id=%u\n", ms->format); ms 142 drivers/gpu/drm/arm/malidp_planes.c drm_printf(p, "\tn_planes=%u\n", ms->n_planes); ms 144 drivers/gpu/drm/arm/malidp_planes.c prefetch_mode_names[ms->mmu_prefetch_mode]); ms 145 drivers/gpu/drm/arm/malidp_planes.c drm_printf(p, "\tmmu_prefetch_pgsize=%d\n", ms->mmu_prefetch_pgsize); ms 332 drivers/gpu/drm/arm/malidp_planes.c static bool malidp_check_pages_threshold(struct malidp_plane_state *ms, ms 337 drivers/gpu/drm/arm/malidp_planes.c for (i = 0; i < ms->n_planes; i++) { ms 343 drivers/gpu/drm/arm/malidp_planes.c obj = drm_gem_fb_get_obj(ms->base.fb, i); ms 436 drivers/gpu/drm/arm/malidp_planes.c (struct malidp_plane_state *ms, u32 *pgsize_bitmap) ms 446 drivers/gpu/drm/arm/malidp_planes.c if (malidp_check_pages_threshold(ms, largest_pgsize)) { ms 457 drivers/gpu/drm/arm/malidp_planes.c if (malidp_partial_prefetch_supported(ms->base.fb->format->format, ms 458 drivers/gpu/drm/arm/malidp_planes.c ms->base.fb->modifier, ms 459 drivers/gpu/drm/arm/malidp_planes.c ms->base.rotation)) { ms 493 drivers/gpu/drm/arm/malidp_planes.c struct malidp_plane_state *ms) ms 499 drivers/gpu/drm/arm/malidp_planes.c ms->mmu_prefetch_pgsize = malidp_get_pgsize_bitmap(mp); ms 500 drivers/gpu/drm/arm/malidp_planes.c ms->mmu_prefetch_mode = ms 501 drivers/gpu/drm/arm/malidp_planes.c malidp_mmu_prefetch_select_mode(ms, &ms->mmu_prefetch_pgsize); ms 508 drivers/gpu/drm/arm/malidp_planes.c struct malidp_plane_state *ms = to_malidp_plane_state(state); ms 520 drivers/gpu/drm/arm/malidp_planes.c ms->format = malidp_hw_get_format_id(&mp->hwdev->hw->map, ms 523 drivers/gpu/drm/arm/malidp_planes.c if (ms->format == MALIDP_INVALID_FORMAT_ID) ms 526 drivers/gpu/drm/arm/malidp_planes.c ms->n_planes = fb->format->num_planes; ms 527 drivers/gpu/drm/arm/malidp_planes.c for (i = 0; i < ms->n_planes; i++) { ms 560 drivers/gpu/drm/arm/malidp_planes.c if (ms->n_planes == 3 && ms 590 drivers/gpu/drm/arm/malidp_planes.c ms->rotmem_size = 0; ms 601 drivers/gpu/drm/arm/malidp_planes.c ms->rotmem_size = val; ms 610 drivers/gpu/drm/arm/malidp_planes.c malidp_de_prefetch_settings(mp, ms); ms 697 drivers/gpu/drm/arm/malidp_planes.c struct malidp_plane_state *ms) ms 705 drivers/gpu/drm/arm/malidp_planes.c mmu_ctrl = malidp_calc_mmu_control_value(ms->mmu_prefetch_mode, ms 707 drivers/gpu/drm/arm/malidp_planes.c ms->n_planes, ms 708 drivers/gpu/drm/arm/malidp_planes.c ms->mmu_prefetch_pgsize); ms 795 drivers/gpu/drm/arm/malidp_planes.c struct malidp_plane_state *ms = to_malidp_plane_state(plane->state); ms 822 drivers/gpu/drm/arm/malidp_planes.c val = (val & ~LAYER_FORMAT_MASK) | ms->format; ms 825 drivers/gpu/drm/arm/malidp_planes.c for (i = 0; i < ms->n_planes; i++) ms 828 drivers/gpu/drm/arm/malidp_planes.c malidp_de_set_mmu_control(mp, ms); ms 830 drivers/gpu/drm/arm/malidp_planes.c malidp_de_set_plane_pitches(mp, ms->n_planes, ms 284 drivers/gpu/drm/i915/selftests/i915_buddy.c u64 s, ms; ms 291 drivers/gpu/drm/i915/selftests/i915_buddy.c ms = BIT_ULL(12 + (prandom_u32_state(&prng) % ilog2(s >> 12))); ms 292 drivers/gpu/drm/i915/selftests/i915_buddy.c s = max(s & -ms, ms); ms 294 drivers/gpu/drm/i915/selftests/i915_buddy.c *chunk_size = ms; ms 241 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c static int msm_hdmi_hdcp_msleep(struct hdmi_hdcp_ctrl *hdcp_ctrl, u32 ms, u32 ev) ms 247 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c msecs_to_jiffies(ms)); ms 126 drivers/gpu/drm/nouveau/nvkm/falcon/base.c nvkm_falcon_wait_for_halt(struct nvkm_falcon *falcon, u32 ms) ms 128 drivers/gpu/drm/nouveau/nvkm/falcon/base.c return falcon->func->wait_for_halt(falcon, ms); ms 292 drivers/gpu/drm/nouveau/nvkm/falcon/v1.c nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms) ms 297 drivers/gpu/drm/nouveau/nvkm/falcon/v1.c ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10); ms 87 drivers/gpu/drm/via/via_video.c (fx->ms / 10) * (HZ / 100), *lock != fx->val); ms 590 drivers/gpu/ipu-v3/ipu-common.c int ipu_idmac_wait_busy(struct ipuv3_channel *channel, int ms) ms 595 drivers/gpu/ipu-v3/ipu-common.c timeout = jiffies + msecs_to_jiffies(ms); ms 62 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 63 drivers/hid/hid-microsoft.c unsigned long quirks = ms->quirks; ms 189 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 190 drivers/hid/hid-microsoft.c unsigned long quirks = ms->quirks; ms 216 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 217 drivers/hid/hid-microsoft.c unsigned long quirks = ms->quirks; ms 228 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 229 drivers/hid/hid-microsoft.c unsigned long quirks = ms->quirks; ms 286 drivers/hid/hid-microsoft.c struct ms_data *ms = container_of(work, struct ms_data, ff_worker); ms 287 drivers/hid/hid-microsoft.c struct hid_device *hdev = ms->hdev; ms 288 drivers/hid/hid-microsoft.c struct xb1s_ff_report *r = ms->output_report_dmabuf; ms 302 drivers/hid/hid-microsoft.c r->magnitude[MAGNITUDE_STRONG] = ms->strong; /* left actuator */ ms 303 drivers/hid/hid-microsoft.c r->magnitude[MAGNITUDE_WEAK] = ms->weak; /* right actuator */ ms 314 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hid); ms 322 drivers/hid/hid-microsoft.c ms->strong = ((u32) effect->u.rumble.strong_magnitude * 100) / U16_MAX; ms 323 drivers/hid/hid-microsoft.c ms->weak = ((u32) effect->u.rumble.weak_magnitude * 100) / U16_MAX; ms 325 drivers/hid/hid-microsoft.c schedule_work(&ms->ff_worker); ms 333 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 342 drivers/hid/hid-microsoft.c if (!(ms->quirks & MS_QUIRK_FF)) ms 345 drivers/hid/hid-microsoft.c ms->hdev = hdev; ms 346 drivers/hid/hid-microsoft.c INIT_WORK(&ms->ff_worker, ms_ff_worker); ms 348 drivers/hid/hid-microsoft.c ms->output_report_dmabuf = devm_kzalloc(&hdev->dev, ms 351 drivers/hid/hid-microsoft.c if (ms->output_report_dmabuf == NULL) ms 360 drivers/hid/hid-microsoft.c struct ms_data *ms = hid_get_drvdata(hdev); ms 362 drivers/hid/hid-microsoft.c if (!(ms->quirks & MS_QUIRK_FF)) ms 365 drivers/hid/hid-microsoft.c cancel_work_sync(&ms->ff_worker); ms 371 drivers/hid/hid-microsoft.c struct ms_data *ms; ms 374 drivers/hid/hid-microsoft.c ms = devm_kzalloc(&hdev->dev, sizeof(*ms), GFP_KERNEL); ms 375 drivers/hid/hid-microsoft.c if (ms == NULL) ms 378 drivers/hid/hid-microsoft.c ms->quirks = quirks; ms 380 drivers/hid/hid-microsoft.c hid_set_drvdata(hdev, ms); ms 614 drivers/hwmon/applesmc.c int ms, ret; ms 616 drivers/hwmon/applesmc.c for (ms = 0; ms < INIT_TIMEOUT_MSECS; ms += INIT_WAIT_MSECS) { ms 619 drivers/hwmon/applesmc.c if (ms) ms 620 drivers/hwmon/applesmc.c pr_info("init_smcreg() took %d ms\n", ms); ms 182 drivers/hwtracing/stm/p_sys-t.c unsigned int ms; ms 186 drivers/hwtracing/stm/p_sys-t.c ret = kstrtouint(page, 10, &ms); ms 190 drivers/hwtracing/stm/p_sys-t.c pn->ts_interval = msecs_to_jiffies(ms); ms 213 drivers/hwtracing/stm/p_sys-t.c unsigned int ms; ms 217 drivers/hwtracing/stm/p_sys-t.c ret = kstrtouint(page, 10, &ms); ms 221 drivers/hwtracing/stm/p_sys-t.c pn->clocksync_interval = msecs_to_jiffies(ms); ms 265 drivers/i2c/busses/i2c-diolan-u2c.c static int diolan_set_clock_synch_timeout(struct i2c_diolan_u2c *dev, int ms) ms 267 drivers/i2c/busses/i2c-diolan-u2c.c int to_val = ms * 10; ms 71 drivers/input/serio/i8042-sparcio.h struct platform_device *ms = of_find_device_by_node(dp); ms 72 drivers/input/serio/i8042-sparcio.h unsigned int irq = ms->archdata.irqs[0]; ms 86 drivers/input/touchscreen/auo-pixcir-ts.c #define AUO_PIXCIR_POWER_IDLE_TIME(ms) ((ms & 0xf) << 4) ms 25 drivers/leds/leds-asic3.c #define MS_TO_CLK(ms) DIV_ROUND_CLOSEST(((ms)*1024), 32000) ms 153 drivers/md/bcache/io.c int ms = us / 1024; ms 157 drivers/md/bcache/io.c ms = min(ms, CONGESTED_MAX + congested); ms 158 drivers/md/bcache/io.c atomic_sub(ms, &c->congested); ms 81 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_gc, sec, ms); ms 83 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_sort, ms, us); ms 84 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_read, ms, us); ms 720 drivers/md/bcache/sysfs.c sysfs_print_time_stats(&c->btree_gc_time, btree_gc, sec, ms); ms 722 drivers/md/bcache/sysfs.c sysfs_print_time_stats(&c->sort.time, btree_sort, ms, us); ms 723 drivers/md/bcache/sysfs.c sysfs_print_time_stats(&c->btree_read_time, btree_read, ms, us); ms 952 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute_list(btree_gc, sec, ms) ms 954 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute_list(btree_sort, ms, us) ms 955 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute_list(btree_read, ms, us) ms 416 drivers/md/dm-integrity.c __u64 ms; ms 419 drivers/md/dm-integrity.c ms = area << ic->sb->log2_interleave_sectors; ms 421 drivers/md/dm-integrity.c ms += area << ic->log2_metadata_run; ms 423 drivers/md/dm-integrity.c ms += area * ic->metadata_run; ms 424 drivers/md/dm-integrity.c ms >>= ic->log2_buffer_sectors; ms 429 drivers/md/dm-integrity.c ms += offset >> (SECTOR_SHIFT + ic->log2_buffer_sectors - ic->log2_tag_size); ms 432 drivers/md/dm-integrity.c ms += (__u64)offset * ic->tag_size >> (SECTOR_SHIFT + ic->log2_buffer_sectors); ms 436 drivers/md/dm-integrity.c return ms; ms 46 drivers/md/dm-raid1.c struct mirror_set *ms; ms 94 drivers/md/dm-raid1.c struct mirror_set *ms = context; ms 96 drivers/md/dm-raid1.c queue_work(ms->kmirrord_wq, &ms->kmirrord_work); ms 101 drivers/md/dm-raid1.c struct mirror_set *ms = from_timer(ms, t, timer); ms 103 drivers/md/dm-raid1.c clear_bit(0, &ms->timer_pending); ms 104 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 107 drivers/md/dm-raid1.c static void delayed_wake(struct mirror_set *ms) ms 109 drivers/md/dm-raid1.c if (test_and_set_bit(0, &ms->timer_pending)) ms 112 drivers/md/dm-raid1.c ms->timer.expires = jiffies + HZ / 5; ms 113 drivers/md/dm-raid1.c add_timer(&ms->timer); ms 121 drivers/md/dm-raid1.c static void queue_bio(struct mirror_set *ms, struct bio *bio, int rw) ms 127 drivers/md/dm-raid1.c bl = (rw == WRITE) ? &ms->writes : &ms->reads; ms 128 drivers/md/dm-raid1.c spin_lock_irqsave(&ms->lock, flags); ms 131 drivers/md/dm-raid1.c spin_unlock_irqrestore(&ms->lock, flags); ms 134 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 139 drivers/md/dm-raid1.c struct mirror_set *ms = context; ms 143 drivers/md/dm-raid1.c queue_bio(ms, bio, WRITE); ms 173 drivers/md/dm-raid1.c static struct mirror *get_default_mirror(struct mirror_set *ms) ms 175 drivers/md/dm-raid1.c return &ms->mirror[atomic_read(&ms->default_mirror)]; ms 180 drivers/md/dm-raid1.c struct mirror_set *ms = m->ms; ms 181 drivers/md/dm-raid1.c struct mirror *m0 = &(ms->mirror[0]); ms 183 drivers/md/dm-raid1.c atomic_set(&ms->default_mirror, m - m0); ms 186 drivers/md/dm-raid1.c static struct mirror *get_valid_mirror(struct mirror_set *ms) ms 190 drivers/md/dm-raid1.c for (m = ms->mirror; m < ms->mirror + ms->nr_mirrors; m++) ms 213 drivers/md/dm-raid1.c struct mirror_set *ms = m->ms; ms 216 drivers/md/dm-raid1.c ms->leg_failure = 1; ms 228 drivers/md/dm-raid1.c if (!errors_handled(ms)) ms 231 drivers/md/dm-raid1.c if (m != get_default_mirror(ms)) ms 234 drivers/md/dm-raid1.c if (!ms->in_sync && !keep_log(ms)) { ms 244 drivers/md/dm-raid1.c new = get_valid_mirror(ms); ms 251 drivers/md/dm-raid1.c schedule_work(&ms->trigger_event); ms 256 drivers/md/dm-raid1.c struct mirror_set *ms = ti->private; ms 267 drivers/md/dm-raid1.c .client = ms->io_client, ms 270 drivers/md/dm-raid1.c for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++) { ms 277 drivers/md/dm-raid1.c dm_io(&io_req, ms->nr_mirrors, io, &error_bits); ms 279 drivers/md/dm-raid1.c for (i = 0; i < ms->nr_mirrors; i++) ms 281 drivers/md/dm-raid1.c fail_mirror(ms->mirror + i, ms 300 drivers/md/dm-raid1.c struct mirror_set *ms = dm_rh_region_context(reg); ms 306 drivers/md/dm-raid1.c fail_mirror(get_default_mirror(ms), DM_RAID1_SYNC_ERROR); ms 316 drivers/md/dm-raid1.c for (m = 0; m < ms->nr_mirrors; m++) { ms 317 drivers/md/dm-raid1.c if (&ms->mirror[m] == get_default_mirror(ms)) ms 320 drivers/md/dm-raid1.c fail_mirror(ms->mirror + m, ms 329 drivers/md/dm-raid1.c static void recover(struct mirror_set *ms, struct dm_region *reg) ms 336 drivers/md/dm-raid1.c sector_t region_size = dm_rh_get_region_size(ms->rh); ms 339 drivers/md/dm-raid1.c m = get_default_mirror(ms); ms 341 drivers/md/dm-raid1.c from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); ms 342 drivers/md/dm-raid1.c if (key == (ms->nr_regions - 1)) { ms 347 drivers/md/dm-raid1.c from.count = ms->ti->len & (region_size - 1); ms 354 drivers/md/dm-raid1.c for (i = 0, dest = to; i < ms->nr_mirrors; i++) { ms 355 drivers/md/dm-raid1.c if (&ms->mirror[i] == get_default_mirror(ms)) ms 358 drivers/md/dm-raid1.c m = ms->mirror + i; ms 360 drivers/md/dm-raid1.c dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); ms 366 drivers/md/dm-raid1.c if (!errors_handled(ms)) ms 369 drivers/md/dm-raid1.c dm_kcopyd_copy(ms->kcopyd_client, &from, ms->nr_mirrors - 1, to, ms 373 drivers/md/dm-raid1.c static void reset_ms_flags(struct mirror_set *ms) ms 377 drivers/md/dm-raid1.c ms->leg_failure = 0; ms 378 drivers/md/dm-raid1.c for (m = 0; m < ms->nr_mirrors; m++) { ms 379 drivers/md/dm-raid1.c atomic_set(&(ms->mirror[m].error_count), 0); ms 380 drivers/md/dm-raid1.c ms->mirror[m].error_type = 0; ms 384 drivers/md/dm-raid1.c static void do_recovery(struct mirror_set *ms) ms 387 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 392 drivers/md/dm-raid1.c dm_rh_recovery_prepare(ms->rh); ms 397 drivers/md/dm-raid1.c while ((reg = dm_rh_recovery_start(ms->rh))) ms 398 drivers/md/dm-raid1.c recover(ms, reg); ms 403 drivers/md/dm-raid1.c if (!ms->in_sync && ms 404 drivers/md/dm-raid1.c (log->type->get_sync_count(log) == ms->nr_regions)) { ms 406 drivers/md/dm-raid1.c dm_table_event(ms->ti->table); ms 407 drivers/md/dm-raid1.c ms->in_sync = 1; ms 408 drivers/md/dm-raid1.c reset_ms_flags(ms); ms 415 drivers/md/dm-raid1.c static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) ms 417 drivers/md/dm-raid1.c struct mirror *m = get_default_mirror(ms); ms 423 drivers/md/dm-raid1.c if (m-- == ms->mirror) ms 424 drivers/md/dm-raid1.c m += ms->nr_mirrors; ms 425 drivers/md/dm-raid1.c } while (m != get_default_mirror(ms)); ms 432 drivers/md/dm-raid1.c struct mirror *default_mirror = get_default_mirror(m->ms); ms 437 drivers/md/dm-raid1.c static int mirror_available(struct mirror_set *ms, struct bio *bio) ms 439 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 440 drivers/md/dm-raid1.c region_t region = dm_rh_bio_to_region(ms->rh, bio); ms 443 drivers/md/dm-raid1.c return choose_mirror(ms, bio->bi_iter.bi_sector) ? 1 : 0; ms 455 drivers/md/dm-raid1.c return m->offset + dm_target_offset(m->ms->ti, bio->bi_iter.bi_sector); ms 472 drivers/md/dm-raid1.c static void hold_bio(struct mirror_set *ms, struct bio *bio) ms 478 drivers/md/dm-raid1.c spin_lock_irq(&ms->lock); ms 480 drivers/md/dm-raid1.c if (atomic_read(&ms->suspend)) { ms 481 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 486 drivers/md/dm-raid1.c if (dm_noflush_suspending(ms->ti)) ms 498 drivers/md/dm-raid1.c bio_list_add(&ms->holds, bio); ms 499 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 520 drivers/md/dm-raid1.c if (likely(default_ok(m)) || mirror_available(m->ms, bio)) { ms 524 drivers/md/dm-raid1.c queue_bio(m->ms, bio, bio_data_dir(bio)); ms 544 drivers/md/dm-raid1.c .client = m->ms->io_client, ms 552 drivers/md/dm-raid1.c static inline int region_in_sync(struct mirror_set *ms, region_t region, ms 555 drivers/md/dm-raid1.c int state = dm_rh_get_state(ms->rh, region, may_block); ms 559 drivers/md/dm-raid1.c static void do_reads(struct mirror_set *ms, struct bio_list *reads) ms 566 drivers/md/dm-raid1.c region = dm_rh_bio_to_region(ms->rh, bio); ms 567 drivers/md/dm-raid1.c m = get_default_mirror(ms); ms 572 drivers/md/dm-raid1.c if (likely(region_in_sync(ms, region, 1))) ms 573 drivers/md/dm-raid1.c m = choose_mirror(ms, bio->bi_iter.bi_sector); ms 600 drivers/md/dm-raid1.c struct mirror_set *ms; ms 604 drivers/md/dm-raid1.c ms = bio_get_m(bio)->ms; ms 628 drivers/md/dm-raid1.c for (i = 0; i < ms->nr_mirrors; i++) ms 630 drivers/md/dm-raid1.c fail_mirror(ms->mirror + i, DM_RAID1_WRITE_ERROR); ms 637 drivers/md/dm-raid1.c spin_lock_irqsave(&ms->lock, flags); ms 638 drivers/md/dm-raid1.c if (!ms->failures.head) ms 640 drivers/md/dm-raid1.c bio_list_add(&ms->failures, bio); ms 641 drivers/md/dm-raid1.c spin_unlock_irqrestore(&ms->lock, flags); ms 643 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 646 drivers/md/dm-raid1.c static void do_write(struct mirror_set *ms, struct bio *bio) ms 658 drivers/md/dm-raid1.c .client = ms->io_client, ms 667 drivers/md/dm-raid1.c for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++) ms 674 drivers/md/dm-raid1.c bio_set_m(bio, get_default_mirror(ms)); ms 676 drivers/md/dm-raid1.c BUG_ON(dm_io(&io_req, ms->nr_mirrors, io, NULL)); ms 679 drivers/md/dm-raid1.c static void do_writes(struct mirror_set *ms, struct bio_list *writes) ms 685 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 706 drivers/md/dm-raid1.c region = dm_rh_bio_to_region(ms->rh, bio); ms 714 drivers/md/dm-raid1.c state = dm_rh_get_state(ms->rh, region, 1); ms 738 drivers/md/dm-raid1.c spin_lock_irq(&ms->lock); ms 739 drivers/md/dm-raid1.c bio_list_merge(&ms->writes, &requeue); ms 740 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 741 drivers/md/dm-raid1.c delayed_wake(ms); ms 749 drivers/md/dm-raid1.c dm_rh_inc_pending(ms->rh, &sync); ms 750 drivers/md/dm-raid1.c dm_rh_inc_pending(ms->rh, &nosync); ms 757 drivers/md/dm-raid1.c ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure; ms 762 drivers/md/dm-raid1.c if (unlikely(ms->log_failure) && errors_handled(ms)) { ms 763 drivers/md/dm-raid1.c spin_lock_irq(&ms->lock); ms 764 drivers/md/dm-raid1.c bio_list_merge(&ms->failures, &sync); ms 765 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 766 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 769 drivers/md/dm-raid1.c do_write(ms, bio); ms 772 drivers/md/dm-raid1.c dm_rh_delay(ms->rh, bio); ms 775 drivers/md/dm-raid1.c if (unlikely(ms->leg_failure) && errors_handled(ms) && !keep_log(ms)) { ms 776 drivers/md/dm-raid1.c spin_lock_irq(&ms->lock); ms 777 drivers/md/dm-raid1.c bio_list_add(&ms->failures, bio); ms 778 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 779 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 781 drivers/md/dm-raid1.c map_bio(get_default_mirror(ms), bio); ms 787 drivers/md/dm-raid1.c static void do_failures(struct mirror_set *ms, struct bio_list *failures) ms 812 drivers/md/dm-raid1.c if (!ms->log_failure) { ms 813 drivers/md/dm-raid1.c ms->in_sync = 0; ms 814 drivers/md/dm-raid1.c dm_rh_mark_nosync(ms->rh, bio); ms 830 drivers/md/dm-raid1.c if (unlikely(!get_valid_mirror(ms) || (keep_log(ms) && ms->log_failure))) ms 832 drivers/md/dm-raid1.c else if (errors_handled(ms) && !keep_log(ms)) ms 833 drivers/md/dm-raid1.c hold_bio(ms, bio); ms 841 drivers/md/dm-raid1.c struct mirror_set *ms = ms 844 drivers/md/dm-raid1.c dm_table_event(ms->ti->table); ms 852 drivers/md/dm-raid1.c struct mirror_set *ms = container_of(work, struct mirror_set, ms 857 drivers/md/dm-raid1.c spin_lock_irqsave(&ms->lock, flags); ms 858 drivers/md/dm-raid1.c reads = ms->reads; ms 859 drivers/md/dm-raid1.c writes = ms->writes; ms 860 drivers/md/dm-raid1.c failures = ms->failures; ms 861 drivers/md/dm-raid1.c bio_list_init(&ms->reads); ms 862 drivers/md/dm-raid1.c bio_list_init(&ms->writes); ms 863 drivers/md/dm-raid1.c bio_list_init(&ms->failures); ms 864 drivers/md/dm-raid1.c spin_unlock_irqrestore(&ms->lock, flags); ms 866 drivers/md/dm-raid1.c dm_rh_update_states(ms->rh, errors_handled(ms)); ms 867 drivers/md/dm-raid1.c do_recovery(ms); ms 868 drivers/md/dm-raid1.c do_reads(ms, &reads); ms 869 drivers/md/dm-raid1.c do_writes(ms, &writes); ms 870 drivers/md/dm-raid1.c do_failures(ms, &failures); ms 881 drivers/md/dm-raid1.c struct mirror_set *ms = ms 882 drivers/md/dm-raid1.c kzalloc(struct_size(ms, mirror, nr_mirrors), GFP_KERNEL); ms 884 drivers/md/dm-raid1.c if (!ms) { ms 889 drivers/md/dm-raid1.c spin_lock_init(&ms->lock); ms 890 drivers/md/dm-raid1.c bio_list_init(&ms->reads); ms 891 drivers/md/dm-raid1.c bio_list_init(&ms->writes); ms 892 drivers/md/dm-raid1.c bio_list_init(&ms->failures); ms 893 drivers/md/dm-raid1.c bio_list_init(&ms->holds); ms 895 drivers/md/dm-raid1.c ms->ti = ti; ms 896 drivers/md/dm-raid1.c ms->nr_mirrors = nr_mirrors; ms 897 drivers/md/dm-raid1.c ms->nr_regions = dm_sector_div_up(ti->len, region_size); ms 898 drivers/md/dm-raid1.c ms->in_sync = 0; ms 899 drivers/md/dm-raid1.c ms->log_failure = 0; ms 900 drivers/md/dm-raid1.c ms->leg_failure = 0; ms 901 drivers/md/dm-raid1.c atomic_set(&ms->suspend, 0); ms 902 drivers/md/dm-raid1.c atomic_set(&ms->default_mirror, DEFAULT_MIRROR); ms 904 drivers/md/dm-raid1.c ms->io_client = dm_io_client_create(); ms 905 drivers/md/dm-raid1.c if (IS_ERR(ms->io_client)) { ms 907 drivers/md/dm-raid1.c kfree(ms); ms 911 drivers/md/dm-raid1.c ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord, ms 913 drivers/md/dm-raid1.c ms->ti->begin, MAX_RECOVERY, ms 914 drivers/md/dm-raid1.c dl, region_size, ms->nr_regions); ms 915 drivers/md/dm-raid1.c if (IS_ERR(ms->rh)) { ms 917 drivers/md/dm-raid1.c dm_io_client_destroy(ms->io_client); ms 918 drivers/md/dm-raid1.c kfree(ms); ms 922 drivers/md/dm-raid1.c return ms; ms 925 drivers/md/dm-raid1.c static void free_context(struct mirror_set *ms, struct dm_target *ti, ms 929 drivers/md/dm-raid1.c dm_put_device(ti, ms->mirror[m].dev); ms 931 drivers/md/dm-raid1.c dm_io_client_destroy(ms->io_client); ms 932 drivers/md/dm-raid1.c dm_region_hash_destroy(ms->rh); ms 933 drivers/md/dm-raid1.c kfree(ms); ms 936 drivers/md/dm-raid1.c static int get_mirror(struct mirror_set *ms, struct dm_target *ti, ms 950 drivers/md/dm-raid1.c &ms->mirror[mirror].dev); ms 956 drivers/md/dm-raid1.c ms->mirror[mirror].ms = ms; ms 957 drivers/md/dm-raid1.c atomic_set(&(ms->mirror[mirror].error_count), 0); ms 958 drivers/md/dm-raid1.c ms->mirror[mirror].error_type = 0; ms 959 drivers/md/dm-raid1.c ms->mirror[mirror].offset = offset; ms 1002 drivers/md/dm-raid1.c static int parse_features(struct mirror_set *ms, unsigned argc, char **argv, ms 1006 drivers/md/dm-raid1.c struct dm_target *ti = ms->ti; ms 1031 drivers/md/dm-raid1.c ms->features |= DM_RAID1_HANDLE_ERRORS; ms 1033 drivers/md/dm-raid1.c ms->features |= DM_RAID1_KEEP_LOG; ms 1043 drivers/md/dm-raid1.c if (!errors_handled(ms) && keep_log(ms)) { ms 1067 drivers/md/dm-raid1.c struct mirror_set *ms; ms 1093 drivers/md/dm-raid1.c ms = alloc_context(nr_mirrors, dl->type->get_region_size(dl), ti, dl); ms 1094 drivers/md/dm-raid1.c if (!ms) { ms 1101 drivers/md/dm-raid1.c r = get_mirror(ms, ti, m, argv); ms 1103 drivers/md/dm-raid1.c free_context(ms, ti, m); ms 1110 drivers/md/dm-raid1.c ti->private = ms; ms 1112 drivers/md/dm-raid1.c r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh)); ms 1120 drivers/md/dm-raid1.c ms->kmirrord_wq = alloc_workqueue("kmirrord", WQ_MEM_RECLAIM, 0); ms 1121 drivers/md/dm-raid1.c if (!ms->kmirrord_wq) { ms 1126 drivers/md/dm-raid1.c INIT_WORK(&ms->kmirrord_work, do_mirror); ms 1127 drivers/md/dm-raid1.c timer_setup(&ms->timer, delayed_wake_fn, 0); ms 1128 drivers/md/dm-raid1.c ms->timer_pending = 0; ms 1129 drivers/md/dm-raid1.c INIT_WORK(&ms->trigger_event, trigger_event); ms 1131 drivers/md/dm-raid1.c r = parse_features(ms, argc, argv, &args_used); ms 1153 drivers/md/dm-raid1.c ms->kcopyd_client = dm_kcopyd_client_create(&dm_kcopyd_throttle); ms 1154 drivers/md/dm-raid1.c if (IS_ERR(ms->kcopyd_client)) { ms 1155 drivers/md/dm-raid1.c r = PTR_ERR(ms->kcopyd_client); ms 1159 drivers/md/dm-raid1.c wakeup_mirrord(ms); ms 1163 drivers/md/dm-raid1.c destroy_workqueue(ms->kmirrord_wq); ms 1165 drivers/md/dm-raid1.c free_context(ms, ti, ms->nr_mirrors); ms 1171 drivers/md/dm-raid1.c struct mirror_set *ms = (struct mirror_set *) ti->private; ms 1173 drivers/md/dm-raid1.c del_timer_sync(&ms->timer); ms 1174 drivers/md/dm-raid1.c flush_workqueue(ms->kmirrord_wq); ms 1175 drivers/md/dm-raid1.c flush_work(&ms->trigger_event); ms 1176 drivers/md/dm-raid1.c dm_kcopyd_client_destroy(ms->kcopyd_client); ms 1177 drivers/md/dm-raid1.c destroy_workqueue(ms->kmirrord_wq); ms 1178 drivers/md/dm-raid1.c free_context(ms, ti, ms->nr_mirrors); ms 1188 drivers/md/dm-raid1.c struct mirror_set *ms = ti->private; ms 1189 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 1197 drivers/md/dm-raid1.c bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio); ms 1198 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); ms 1202 drivers/md/dm-raid1.c r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); ms 1213 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); ms 1221 drivers/md/dm-raid1.c m = choose_mirror(ms, bio->bi_iter.bi_sector); ms 1237 drivers/md/dm-raid1.c struct mirror_set *ms = (struct mirror_set *) ti->private; ms 1249 drivers/md/dm-raid1.c dm_rh_dec(ms->rh, bio_record->write_region); ms 1281 drivers/md/dm-raid1.c if (default_ok(m) || mirror_available(ms, bio)) { ms 1288 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); ms 1302 drivers/md/dm-raid1.c struct mirror_set *ms = (struct mirror_set *) ti->private; ms 1303 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 1308 drivers/md/dm-raid1.c atomic_set(&ms->suspend, 1); ms 1316 drivers/md/dm-raid1.c spin_lock_irq(&ms->lock); ms 1317 drivers/md/dm-raid1.c holds = ms->holds; ms 1318 drivers/md/dm-raid1.c bio_list_init(&ms->holds); ms 1319 drivers/md/dm-raid1.c spin_unlock_irq(&ms->lock); ms 1322 drivers/md/dm-raid1.c hold_bio(ms, bio); ms 1328 drivers/md/dm-raid1.c dm_rh_stop_recovery(ms->rh); ms 1331 drivers/md/dm-raid1.c !dm_rh_recovery_in_flight(ms->rh)); ms 1343 drivers/md/dm-raid1.c flush_workqueue(ms->kmirrord_wq); ms 1348 drivers/md/dm-raid1.c struct mirror_set *ms = ti->private; ms 1349 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 1358 drivers/md/dm-raid1.c struct mirror_set *ms = ti->private; ms 1359 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 1361 drivers/md/dm-raid1.c atomic_set(&ms->suspend, 0); ms 1365 drivers/md/dm-raid1.c dm_rh_start_recovery(ms->rh); ms 1398 drivers/md/dm-raid1.c struct mirror_set *ms = (struct mirror_set *) ti->private; ms 1399 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); ms 1404 drivers/md/dm-raid1.c DMEMIT("%d ", ms->nr_mirrors); ms 1405 drivers/md/dm-raid1.c for (m = 0; m < ms->nr_mirrors; m++) { ms 1406 drivers/md/dm-raid1.c DMEMIT("%s ", ms->mirror[m].dev->name); ms 1407 drivers/md/dm-raid1.c buffer[m] = device_status_char(&(ms->mirror[m])); ms 1413 drivers/md/dm-raid1.c (unsigned long long)ms->nr_regions, buffer); ms 1422 drivers/md/dm-raid1.c DMEMIT("%d", ms->nr_mirrors); ms 1423 drivers/md/dm-raid1.c for (m = 0; m < ms->nr_mirrors; m++) ms 1424 drivers/md/dm-raid1.c DMEMIT(" %s %llu", ms->mirror[m].dev->name, ms 1425 drivers/md/dm-raid1.c (unsigned long long)ms->mirror[m].offset); ms 1427 drivers/md/dm-raid1.c num_feature_args += !!errors_handled(ms); ms 1428 drivers/md/dm-raid1.c num_feature_args += !!keep_log(ms); ms 1431 drivers/md/dm-raid1.c if (errors_handled(ms)) ms 1433 drivers/md/dm-raid1.c if (keep_log(ms)) ms 1444 drivers/md/dm-raid1.c struct mirror_set *ms = ti->private; ms 1448 drivers/md/dm-raid1.c for (i = 0; !ret && i < ms->nr_mirrors; i++) ms 1449 drivers/md/dm-raid1.c ret = fn(ti, ms->mirror[i].dev, ms 1450 drivers/md/dm-raid1.c ms->mirror[i].offset, ti->len, data); ms 173 drivers/media/dvb-frontends/horus3a.c u32 ms = 0; ms 196 drivers/media/dvb-frontends/horus3a.c ms = DIV_ROUND_CLOSEST((frequency * mixdiv) / 2, 1000); ms 197 drivers/media/dvb-frontends/horus3a.c if (ms > 0x7FFF) { /* 15 bit */ ms 292 drivers/media/dvb-frontends/horus3a.c data[0] = (u8)((ms >> 7) & 0xFF); ms 293 drivers/media/dvb-frontends/horus3a.c data[1] = (u8)((ms << 1) & 0xFF); ms 309 drivers/media/dvb-frontends/horus3a.c priv->frequency = ms * 2 * 1000 / mixdiv; ms 300 drivers/media/i2c/m5mols/m5mols_core.c int ms = timeout < 0 ? M5MOLS_BUSY_WAIT_DEF_TIMEOUT : timeout; ms 301 drivers/media/i2c/m5mols/m5mols_core.c unsigned long end = jiffies + msecs_to_jiffies(ms); ms 312 drivers/media/i2c/m5mols/m5mols_core.c } while (ms > 0 && time_is_after_jiffies(end)); ms 278 drivers/media/pci/solo6x10/solo6x10-core.c unsigned long ms; ms 279 drivers/media/pci/solo6x10/solo6x10-core.c int ret = kstrtoul(buf, 10, &ms); ms 281 drivers/media/pci/solo6x10/solo6x10-core.c if (ret < 0 || ms > 200) ms 283 drivers/media/pci/solo6x10/solo6x10-core.c solo_dev->p2m_jiffies = msecs_to_jiffies(ms); ms 415 drivers/media/platform/vivid/vivid-kthread-cap.c unsigned ms; ms 478 drivers/media/platform/vivid/vivid-kthread-cap.c ms = dev->ms_vid_cap; ms 481 drivers/media/platform/vivid/vivid-kthread-cap.c (ms / (60 * 60 * 1000)) % 24, ms 482 drivers/media/platform/vivid/vivid-kthread-cap.c (ms / (60 * 1000)) % 60, ms 483 drivers/media/platform/vivid/vivid-kthread-cap.c (ms / 1000) % 60, ms 484 drivers/media/platform/vivid/vivid-kthread-cap.c ms % 1000, ms 78 drivers/media/platform/vivid/vivid-radio-common.c rds->ms = dev->radio_tx_rds_ms->cur.val; ms 93 drivers/media/platform/vivid/vivid-radio-common.c v4l2_ctrl_s_ctrl(dev->radio_rx_rds_ms, rds->ms); ms 66 drivers/media/platform/vivid/vivid-rds-gen.c data[1].lsb |= (rds->ta << 4) | (rds->ms << 3); ms 113 drivers/media/platform/vivid/vivid-rds-gen.c data[1].lsb |= (rds->ta << 4) | (rds->ms << 3); ms 120 drivers/media/platform/vivid/vivid-rds-gen.c data[3].lsb |= (rds->ta << 4) | (rds->ms << 3); ms 146 drivers/media/platform/vivid/vivid-rds-gen.c rds->ms = true; ms 33 drivers/media/platform/vivid/vivid-rds-gen.h bool ms; ms 977 drivers/mmc/core/block.c unsigned int ms = DIV_ROUND_UP(data->timeout_ns, 1000000); ms 982 drivers/mmc/core/block.c ms += DIV_ROUND_UP(data->timeout_clks, khz); ms 985 drivers/mmc/core/block.c return ms; ms 61 drivers/mmc/core/core.h static inline void mmc_delay(unsigned int ms) ms 63 drivers/mmc/core/core.h if (ms <= 20) ms 64 drivers/mmc/core/core.h usleep_range(ms * 1000, ms * 1250); ms 66 drivers/mmc/core/core.h msleep(ms); ms 336 drivers/mmc/host/mxs-mmc.c const unsigned int ms = ns / 1000; ms 337 drivers/mmc/host/mxs-mmc.c const unsigned int ticks = ms * clock_per_ms; ms 164 drivers/mtd/tests/speedtest.c long ms; ms 166 drivers/mtd/tests/speedtest.c ms = ktime_ms_delta(finish, start); ms 167 drivers/mtd/tests/speedtest.c if (ms == 0) ms 170 drivers/mtd/tests/speedtest.c do_div(k, ms); ms 322 drivers/mtd/tests/torturetest.c long ms; ms 325 drivers/mtd/tests/torturetest.c ms = ktime_ms_delta(finish, start); ms 328 drivers/mtd/tests/torturetest.c erase_cycles, ms, ms / 1000); ms 21 drivers/net/dsa/sja1105/sja1105.h #define SJA1105_AGEING_TIME_MS(ms) ((ms) / 10) ms 2063 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h static inline u32 reg_poll(struct bnx2x *bp, u32 reg, u32 expected, int ms, ms 2072 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h ms -= wait; ms 2075 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h } while (ms > 0); ms 1400 drivers/net/ethernet/cavium/liquidio/octeon_device.c u32 ms; ms 1405 drivers/net/ethernet/cavium/liquidio/octeon_device.c for (ms = 0; (ret != 0) && ((*timeout == 0) || (ms <= *timeout)); ms 1406 drivers/net/ethernet/cavium/liquidio/octeon_device.c ms += HZ / 10) { ms 294 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int i, ms, delay_idx, ret; ms 327 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ms = delay[0]; ms 329 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c for (i = 0; ; i += ms) { ms 354 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ms = delay[delay_idx]; /* last element may repeat */ ms 357 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c msleep(ms); ms 359 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c mdelay(ms); ms 387 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ms = delay[0]; ms 392 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c i += ms) { ms 394 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ms = delay[delay_idx]; /* last element may repeat */ ms 397 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c msleep(ms); ms 399 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c mdelay(ms); ms 420 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c execute = i + ms; ms 7131 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int ms; ms 7134 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c for (ms = 0; ms < FW_CMD_MAX_TIMEOUT; ) { ms 7138 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ms += 100; ms 137 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c int i, ms, delay_idx, ret; ms 170 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c ms = delay[0]; ms 172 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c for (i = 0; ; i += ms) { ms 196 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c ms = delay[delay_idx]; /* last element may repeat */ ms 199 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c msleep(ms); ms 201 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c mdelay(ms); ms 248 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c ms = delay[0]; ms 250 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c for (i = 0; i < FW_CMD_MAX_TIMEOUT; i += ms) { ms 252 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c ms = delay[delay_idx]; ms 255 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c msleep(ms); ms 257 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c mdelay(ms); ms 296 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c execute = i + ms; ms 268 drivers/net/ethernet/davicom/dm9000.c static void dm9000_msleep(struct board_info *db, unsigned int ms) ms 271 drivers/net/ethernet/davicom/dm9000.c mdelay(ms); ms 273 drivers/net/ethernet/davicom/dm9000.c msleep(ms); ms 1303 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_ms_reg_ctrl *ms) ms 1305 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->control = QLCNIC_MS_CTRL; ms 1306 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->low = QLCNIC_MS_ADDR_LO; ms 1307 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->hi = QLCNIC_MS_ADDR_HI; ms 1309 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[0] = QLCNIC_MS_WRTDATA_LO; ms 1310 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[0] = QLCNIC_MS_RDDATA_LO; ms 1311 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[1] = QLCNIC_MS_WRTDATA_HI; ms 1312 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[1] = QLCNIC_MS_RDDATA_HI; ms 1313 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[2] = QLCNIC_MS_WRTDATA_ULO; ms 1314 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[3] = QLCNIC_MS_WRTDATA_UHI; ms 1315 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[2] = QLCNIC_MS_RDDATA_ULO; ms 1316 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[3] = QLCNIC_MS_RDDATA_UHI; ms 1318 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[0] = QLCNIC_MS_WRTDATA_ULO; ms 1319 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[0] = QLCNIC_MS_RDDATA_ULO; ms 1320 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[1] = QLCNIC_MS_WRTDATA_UHI; ms 1321 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[1] = QLCNIC_MS_RDDATA_UHI; ms 1322 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[2] = QLCNIC_MS_WRTDATA_LO; ms 1323 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->wd[3] = QLCNIC_MS_WRTDATA_HI; ms 1324 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[2] = QLCNIC_MS_RDDATA_LO; ms 1325 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->rd[3] = QLCNIC_MS_RDDATA_HI; ms 1328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->ocm_window = OCM_WIN_P3P(off); ms 1329 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms->off = GET_MEM_OFFS_2M(off); ms 1336 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_ms_reg_ctrl ms; ms 1342 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl)); ms 1349 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_set_ms_controls(adapter, off, &ms); ms 1352 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window, ms 1353 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms.off, &data, 1); ms 1359 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.low, off8); ms 1360 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.hi, 0); ms 1362 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE); ms 1363 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE); ms 1366 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c temp = qlcnic_ind_rd(adapter, ms.control); ms 1377 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.wd[0], qlcnic_ind_rd(adapter, ms.rd[0])); ms 1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.wd[1], qlcnic_ind_rd(adapter, ms.rd[1])); ms 1380 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.wd[2], data & 0xffffffff); ms 1381 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.wd[3], (data >> 32) & 0xffffffff); ms 1383 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_ENABLE); ms 1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_START); ms 1387 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c temp = qlcnic_ind_rd(adapter, ms.control); ms 1411 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_ms_reg_ctrl ms; ms 1422 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl)); ms 1423 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_set_ms_controls(adapter, off, &ms); ms 1426 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window, ms 1427 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c ms.off, data, 0); ms 1433 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.low, off8); ms 1434 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.hi, 0); ms 1436 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE); ms 1437 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE); ms 1440 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c temp = qlcnic_ind_rd(adapter, ms.control); ms 1452 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c temp = qlcnic_ind_rd(adapter, ms.rd[3]); ms 1454 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c val |= qlcnic_ind_rd(adapter, ms.rd[2]); ms 45 drivers/net/fddi/defxx.h PI_UINT32 ms; ms 280 drivers/net/hamradio/yam.c static void delay(int ms) ms 282 drivers/net/hamradio/yam.c unsigned long timeout = jiffies + ((ms * HZ) / 1000); ms 669 drivers/net/ieee802154/ca8210.c static void ca8210_reset_send(struct spi_device *spi, unsigned int ms) ms 677 drivers/net/ieee802154/ca8210.c msleep(ms); ms 389 drivers/net/wimax/i2400m/control.c const struct i2400m_tlv_media_status *ms) ms 394 drivers/net/wimax/i2400m/control.c enum i2400m_media_status status = le32_to_cpu(ms->media_status); ms 396 drivers/net/wimax/i2400m/control.c d_fnstart(3, dev, "(i2400m %p ms %p [%u])\n", i2400m, ms, status); ms 418 drivers/net/wimax/i2400m/control.c i2400m, ms, status); ms 438 drivers/net/wimax/i2400m/control.c const struct i2400m_tlv_media_status *ms; ms 459 drivers/net/wimax/i2400m/control.c if (0 == i2400m_tlv_match(tlv, I2400M_TLV_MEDIA_STATUS, sizeof(*ms))) { ms 460 drivers/net/wimax/i2400m/control.c ms = container_of(tlv, typeof(*ms), hdr); ms 462 drivers/net/wimax/i2400m/control.c tag, le32_to_cpu(ms->media_status)); ms 463 drivers/net/wimax/i2400m/control.c i2400m_report_tlv_media_status(i2400m, ms); ms 955 drivers/net/wimax/i2400m/i2400m.h void __i2400m_msleep(unsigned ms) ms 959 drivers/net/wimax/i2400m/i2400m.h msleep(ms); ms 235 drivers/net/wireless/ath/ath9k/channel.c u64 ms; ms 240 drivers/net/wireless/ath/ath9k/channel.c ms = ts.tv_sec * 1000 + ts.tv_nsec / 1000000; ms 241 drivers/net/wireless/ath/ath9k/channel.c ms -= old->tv_sec * 1000 + old->tv_nsec / 1000000; ms 244 drivers/net/wireless/ath/ath9k/channel.c return (u32)ms; ms 620 drivers/net/wireless/ath/carl9170/rx.c static bool carl9170_ampdu_check(struct ar9170 *ar, u8 *buf, u8 ms, ms 625 drivers/net/wireless/ath/carl9170/rx.c if ((ms & AR9170_RX_STATUS_MPDU) == AR9170_RX_STATUS_MPDU_SINGLE) { ms 2502 drivers/net/wireless/ath/wil6210/cfg80211.c struct cfg80211_match_set *ms = &request->match_sets[i]; ms 2505 drivers/net/wireless/ath/wil6210/cfg80211.c i, ms->rssi_thold); ms 2507 drivers/net/wireless/ath/wil6210/cfg80211.c ms->ssid.ssid, ms 2508 drivers/net/wireless/ath/wil6210/cfg80211.c ms->ssid.ssid_len, true); ms 3465 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c static __always_inline void brcmf_delay(u32 ms) ms 3467 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c if (ms < 1000 / HZ) { ms 3469 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c mdelay(ms); ms 3471 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c msleep(ms); ms 363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c struct cfg80211_match_set *ms; ms 371 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c ms = &r->match_sets[j]; ms 372 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c if (ms->ssid.ssid_len) { ms 373 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c active = brcmf_is_ssid_active(&ms->ssid, r); ms 374 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c err = brcmf_pno_add_ssid(ifp, &ms->ssid, ms 377 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c if (!err && is_valid_ether_addr(ms->bssid)) ms 378 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c err = brcmf_pno_add_bssid(ifp, ms->bssid); ms 565 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c struct cfg80211_match_set *ms; ms 576 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c ms = &req->match_sets[j]; ms 577 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c if (ms->ssid.ssid_len == ni->SSID_len && ms 578 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c !memcmp(ms->ssid.ssid, ni->SSID, ni->SSID_len)) { ms 582 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c if (is_valid_ether_addr(ms->bssid) && ms 583 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c !memcmp(ms->bssid, ni->bssid, ETH_ALEN)) { ms 43 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h __field(uint, ms) ms 48 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h __entry->ms = t->ms; ms 54 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h __entry->ms, __entry->set, __entry->periodic ms 1468 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c msecs_to_jiffies(t->ms)); ms 1517 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c void brcms_add_timer(struct brcms_timer *t, uint ms, int periodic) ms 1527 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c t->ms = ms; ms 1534 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c ieee80211_queue_delayed_work(hw, &t->dly_wrk, msecs_to_jiffies(ms)); ms 41 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h uint ms; ms 107 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h void brcms_add_timer(struct brcms_timer *timer, uint ms, int periodic); ms 74 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic) ms 76 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c brcms_add_timer((struct brcms_timer *)t, ms, periodic); ms 137 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h void wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic); ms 1364 drivers/net/wireless/intel/iwlegacy/common.c il_scan_cancel_timeout(struct il_priv *il, unsigned long ms) ms 1366 drivers/net/wireless/intel/iwlegacy/common.c unsigned long timeout = jiffies + msecs_to_jiffies(ms); ms 1773 drivers/net/wireless/intel/iwlegacy/common.h int il_scan_cancel_timeout(struct il_priv *il, unsigned long ms); ms 247 drivers/net/wireless/intel/iwlwifi/dvm/agn.h void iwl_scan_cancel_timeout(struct iwl_priv *priv, unsigned long ms); ms 204 drivers/net/wireless/intel/iwlwifi/dvm/scan.c void iwl_scan_cancel_timeout(struct iwl_priv *priv, unsigned long ms) ms 206 drivers/net/wireless/intel/iwlwifi/dvm/scan.c unsigned long timeout = jiffies + msecs_to_jiffies(ms); ms 2269 drivers/net/wireless/marvell/mwl8k.c int ms; ms 2271 drivers/net/wireless/marvell/mwl8k.c ms = MWL8K_CMD_TIMEOUT_MS - jiffies_to_msecs(timeout); ms 2278 drivers/net/wireless/marvell/mwl8k.c else if (ms > 2000) ms 2282 drivers/net/wireless/marvell/mwl8k.c ms); ms 215 drivers/nvme/host/core.c return ns->pi_type && ns->ms == sizeof(struct t10_pi_tuple); ms 707 drivers/nvme/host/core.c if (ns->ms) { ms 1291 drivers/nvme/host/core.c meta_len = (io.nblocks + 1) * ns->ms; ms 1645 drivers/nvme/host/core.c static void nvme_init_integrity(struct gendisk *disk, u16 ms, u8 pi_type) ms 1666 drivers/nvme/host/core.c integrity.tuple_size = ms; ms 1671 drivers/nvme/host/core.c static void nvme_init_integrity(struct gendisk *disk, u16 ms, u8 pi_type) ms 1813 drivers/nvme/host/core.c if (ns->ms && !ns->ext && ms 1815 drivers/nvme/host/core.c nvme_init_integrity(disk, ns->ms, ns->pi_type); ms 1816 drivers/nvme/host/core.c if ((ns->ms && !nvme_ns_has_pi(ns) && !blk_get_integrity(disk)) || ms 1845 drivers/nvme/host/core.c ns->ms = le16_to_cpu(id->lbaf[id->flbas & NVME_NS_FLBAS_LBA_MASK].ms); ms 1846 drivers/nvme/host/core.c ns->ext = ns->ms && (id->flbas & NVME_NS_FLBAS_META_EXT); ms 1848 drivers/nvme/host/core.c if (ns->ms == sizeof(struct t10_pi_tuple)) ms 963 drivers/nvme/host/lightnvm.c geo->sos = ns->ms; ms 368 drivers/nvme/host/nvme.h u16 ms; ms 880 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c DB8500_FUNC_GROUPS(ms, "ms_c_1"); ms 939 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c FUNCTION(ms), ms 55 drivers/pwm/pwm-lpss.c const unsigned int ms = 500 * USEC_PER_MSEC; ms 70 drivers/pwm/pwm-lpss.c err = readl_poll_timeout(addr, val, !(val & PWM_SW_UPDATE), 40, ms); ms 2319 drivers/regulator/core.c unsigned int ms = delay / 1000; ms 2322 drivers/regulator/core.c if (ms > 0) { ms 2327 drivers/regulator/core.c if (ms < 20) ms 2328 drivers/regulator/core.c us += ms * 1000; ms 2330 drivers/regulator/core.c msleep(ms); ms 2784 drivers/regulator/core.c int regulator_disable_deferred(struct regulator *regulator, int ms) ms 2788 drivers/regulator/core.c if (!ms) ms 2794 drivers/regulator/core.c msecs_to_jiffies(ms)); ms 421 drivers/remoteproc/qcom_q6v5_mss.c static int q6v5_rmb_pbl_wait(struct q6v5 *qproc, int ms) ms 426 drivers/remoteproc/qcom_q6v5_mss.c timeout = jiffies + msecs_to_jiffies(ms); ms 441 drivers/remoteproc/qcom_q6v5_mss.c static int q6v5_rmb_mba_wait(struct q6v5 *qproc, u32 status, int ms) ms 447 drivers/remoteproc/qcom_q6v5_mss.c timeout = jiffies + msecs_to_jiffies(ms); ms 225 drivers/s390/scsi/zfcp_fc.c zfcp_fc_wka_port_force_offline(&gs->ms); ms 977 drivers/s390/scsi/zfcp_fc.c return &adapter->gs->ms; ms 1091 drivers/s390/scsi/zfcp_fc.c zfcp_fc_wka_port_init(&wka_ports->ms, FC_FID_MGMT_SERV, adapter); ms 215 drivers/s390/scsi/zfcp_fc.h struct zfcp_fc_wka_port ms; ms 665 drivers/scsi/aic94xx/aic94xx_sds.c static int asd_validate_ms(struct asd_manuf_sec *ms) ms 667 drivers/scsi/aic94xx/aic94xx_sds.c if (ms->sig[0] != 'S' || ms->sig[1] != 'M') { ms 669 drivers/scsi/aic94xx/aic94xx_sds.c ms->sig[0], ms->sig[1]); ms 672 drivers/scsi/aic94xx/aic94xx_sds.c if (ms->maj != 0) { ms 674 drivers/scsi/aic94xx/aic94xx_sds.c ms->maj); ms 677 drivers/scsi/aic94xx/aic94xx_sds.c ms->offs_next = le16_to_cpu((__force __le16) ms->offs_next); ms 678 drivers/scsi/aic94xx/aic94xx_sds.c ms->chksum = le16_to_cpu((__force __le16) ms->chksum); ms 679 drivers/scsi/aic94xx/aic94xx_sds.c ms->size = le16_to_cpu((__force __le16) ms->size); ms 681 drivers/scsi/aic94xx/aic94xx_sds.c if (asd_calc_flash_chksum((u16 *)ms, ms->size/2)) { ms 689 drivers/scsi/aic94xx/aic94xx_sds.c struct asd_manuf_sec *ms) ms 691 drivers/scsi/aic94xx/aic94xx_sds.c memcpy(asd_ha->hw_prof.sas_addr, ms->sas_addr, SAS_ADDR_SIZE); ms 696 drivers/scsi/aic94xx/aic94xx_sds.c struct asd_manuf_sec *ms) ms 698 drivers/scsi/aic94xx/aic94xx_sds.c memcpy(asd_ha->hw_prof.pcba_sn, ms->pcba_sn, ASD_PCBA_SN_SIZE); ms 85 drivers/scsi/bfa/bfa_fcs.h struct bfa_fcs_lport_ms_s *ms; /* parent ms */ ms 108 drivers/scsi/bfa/bfa_fcs.h struct bfa_fcs_lport_ms_s ms; /* MS component of port */ ms 157 drivers/scsi/bfa/bfa_fcs.h #define BFA_FCS_GET_MS_FROM_PORT(port) (&port->port_topo.pfab.ms) ms 158 drivers/scsi/bfa/bfa_fcs.h #define BFA_FCS_GET_FDMI_FROM_PORT(port) (&port->port_topo.pfab.ms.fdmi) ms 310 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_init(struct bfa_fcs_lport_ms_s *ms); ms 311 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_offline(struct bfa_fcs_lport_ms_s *ms); ms 312 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_online(struct bfa_fcs_lport_ms_s *ms); ms 1481 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1521 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1546 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1595 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1626 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1651 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1698 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1729 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1754 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1799 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1827 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1848 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1863 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 1902 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2162 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2196 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2462 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2485 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2519 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2556 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2577 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2616 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2672 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = fdmi->ms->port; ms 2791 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_init(struct bfa_fcs_lport_ms_s *ms) ms 2793 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; ms 2795 drivers/scsi/bfa/bfa_fcs_lport.c fdmi->ms = ms; ms 2796 drivers/scsi/bfa/bfa_fcs_lport.c if (ms->port->fcs->fdmi_enabled) ms 2803 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_offline(struct bfa_fcs_lport_ms_s *ms) ms 2805 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; ms 2807 drivers/scsi/bfa/bfa_fcs_lport.c fdmi->ms = ms; ms 2812 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_online(struct bfa_fcs_lport_ms_s *ms) ms 2814 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; ms 2816 drivers/scsi/bfa/bfa_fcs_lport.c fdmi->ms = ms; ms 2871 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms, ms 2873 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms, ms 2875 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms, ms 2877 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms, ms 2879 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms, ms 2881 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms, ms 2883 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms, ms 2885 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms, ms 2887 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms, ms 2889 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms, ms 2891 drivers/scsi/bfa/bfa_fcs_lport.c static void bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms, ms 2897 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms, ms 2900 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 2901 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 2905 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending); ms 2906 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_plogi(ms, NULL); ms 2913 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 2918 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms, ms 2921 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 2922 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 2926 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi); ms 2930 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 2931 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 2932 drivers/scsi/bfa/bfa_fcs_lport.c &ms->fcxp_wqe); ms 2936 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 2941 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms, ms 2944 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 2945 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 2952 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_retry); ms 2953 drivers/scsi/bfa/bfa_fcs_lport.c ms->port->stats.ms_retries++; ms 2954 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 2955 drivers/scsi/bfa/bfa_fcs_lport.c &ms->timer, bfa_fcs_lport_ms_timeout, ms, ms 2963 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_online(ms); ms 2968 drivers/scsi/bfa/bfa_fcs_lport.c if (ms->port->vport) { ms 2969 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); ms 2977 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending); ms 2978 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gmal(ms, NULL); ms 2982 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 2983 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_discard(ms->fcxp); ms 2987 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 2992 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms, ms 2995 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 2996 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3003 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending); ms 3004 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_plogi(ms, NULL); ms 3008 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3009 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_stop(&ms->timer); ms 3013 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3018 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms, ms 3021 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3022 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3026 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3030 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); ms 3031 drivers/scsi/bfa/bfa_fcs_lport.c ms->retry_cnt = 0; ms 3032 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gfn(ms, NULL); ms 3036 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3041 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms, ms 3044 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3045 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3049 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal); ms 3053 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3054 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 3055 drivers/scsi/bfa/bfa_fcs_lport.c &ms->fcxp_wqe); ms 3059 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3064 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms, ms 3067 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3068 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3075 drivers/scsi/bfa/bfa_fcs_lport.c if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) { ms 3076 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_retry); ms 3077 drivers/scsi/bfa/bfa_fcs_lport.c ms->port->stats.ms_retries++; ms 3078 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 3079 drivers/scsi/bfa/bfa_fcs_lport.c &ms->timer, bfa_fcs_lport_ms_timeout, ms, ms 3082 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); ms 3083 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gfn(ms, NULL); ms 3084 drivers/scsi/bfa/bfa_fcs_lport.c ms->retry_cnt = 0; ms 3089 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); ms 3090 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gfn(ms, NULL); ms 3094 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3095 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_discard(ms->fcxp); ms 3099 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3104 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms, ms 3107 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3108 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3115 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending); ms 3116 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gmal(ms, NULL); ms 3120 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3121 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_stop(&ms->timer); ms 3125 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3135 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = ms_cbarg; ms 3136 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_t *port = ms->port; ms 3146 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, ms 3147 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gmal, ms, BFA_TRUE); ms 3150 drivers/scsi/bfa/bfa_fcs_lport.c ms->fcxp = fcxp; ms 3158 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_gmal_response, (void *)ms, ms 3161 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); ms 3170 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; ms 3171 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_t *port = ms->port; ms 3186 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3198 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3234 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); ms 3240 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3244 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms, ms 3247 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3248 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3252 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn); ms 3256 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3257 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 3258 drivers/scsi/bfa/bfa_fcs_lport.c &ms->fcxp_wqe); ms 3262 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3267 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms, ms 3270 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3271 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3278 drivers/scsi/bfa/bfa_fcs_lport.c if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) { ms 3279 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_retry); ms 3280 drivers/scsi/bfa/bfa_fcs_lport.c ms->port->stats.ms_retries++; ms 3281 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), ms 3282 drivers/scsi/bfa/bfa_fcs_lport.c &ms->timer, bfa_fcs_lport_ms_timeout, ms, ms 3285 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); ms 3286 drivers/scsi/bfa/bfa_fcs_lport.c ms->retry_cnt = 0; ms 3291 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); ms 3295 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3296 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcxp_discard(ms->fcxp); ms 3300 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3305 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms, ms 3308 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); ms 3309 drivers/scsi/bfa/bfa_fcs_lport.c bfa_trc(ms->port->fcs, event); ms 3316 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); ms 3317 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gfn(ms, NULL); ms 3321 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3322 drivers/scsi/bfa/bfa_fcs_lport.c bfa_timer_stop(&ms->timer); ms 3326 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_fault(ms->port->fcs, event); ms 3336 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = ms_cbarg; ms 3337 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_t *port = ms->port; ms 3347 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, ms 3348 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_gfn, ms, BFA_TRUE); ms 3351 drivers/scsi/bfa/bfa_fcs_lport.c ms->fcxp = fcxp; ms 3359 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_gfn_response, (void *)ms, ms 3362 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); ms 3370 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; ms 3371 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_t *port = ms->port; ms 3383 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3397 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); ms 3403 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3413 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = ms_cbarg; ms 3414 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = ms->port; ms 3425 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, ms 3426 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_send_plogi, ms, BFA_TRUE); ms 3429 drivers/scsi/bfa/bfa_fcs_lport.c ms->fcxp = fcxp; ms 3440 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_plogi_response, (void *)ms, ms 3444 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); ms 3452 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; ms 3453 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_s *port = ms->port; ms 3466 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3478 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3482 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); ms 3492 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3498 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); ms 3505 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) arg; ms 3507 drivers/scsi/bfa/bfa_fcs_lport.c ms->port->stats.ms_timeouts++; ms 3508 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_TIMEOUT); ms 3515 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); ms 3517 drivers/scsi/bfa/bfa_fcs_lport.c ms->port = port; ms 3518 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); ms 3523 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_init(ms); ms 3529 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); ms 3531 drivers/scsi/bfa/bfa_fcs_lport.c ms->port = port; ms 3532 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_PORT_OFFLINE); ms 3533 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_offline(ms); ms 3539 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); ms 3541 drivers/scsi/bfa/bfa_fcs_lport.c ms->port = port; ms 3542 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_PORT_ONLINE); ms 3547 drivers/scsi/bfa/bfa_fcs_lport.c struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); ms 3550 drivers/scsi/bfa/bfa_fcs_lport.c if (bfa_sm_cmp_state(ms, bfa_fcs_lport_ms_sm_online)) ms 3551 drivers/scsi/bfa/bfa_fcs_lport.c bfa_sm_send_event(ms, MSSM_EVENT_PORT_FABRIC_RSCN); ms 1286 drivers/scsi/csiostor/csio_hw.c int ms; ms 1289 drivers/scsi/csiostor/csio_hw.c for (ms = 0; ms < FW_CMD_MAX_TIMEOUT; ) { ms 1293 drivers/scsi/csiostor/csio_hw.c ms += 100; ms 151 drivers/scsi/lpfc/lpfc_debugfs.c uint32_t ms; ms 169 drivers/scsi/lpfc/lpfc_debugfs.c ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time); ms 172 drivers/scsi/lpfc/lpfc_debugfs.c dtp->seq_cnt, ms, dtp->fmt); ms 180 drivers/scsi/lpfc/lpfc_debugfs.c ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time); ms 183 drivers/scsi/lpfc/lpfc_debugfs.c dtp->seq_cnt, ms, dtp->fmt); ms 217 drivers/scsi/lpfc/lpfc_debugfs.c uint32_t ms; ms 235 drivers/scsi/lpfc/lpfc_debugfs.c ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time); ms 238 drivers/scsi/lpfc/lpfc_debugfs.c dtp->seq_cnt, ms, dtp->fmt); ms 246 drivers/scsi/lpfc/lpfc_debugfs.c ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time); ms 249 drivers/scsi/lpfc/lpfc_debugfs.c dtp->seq_cnt, ms, dtp->fmt); ms 186 drivers/scsi/mesh.c static void mesh_done(struct mesh_state *ms, int start_next); ms 187 drivers/scsi/mesh.c static void mesh_interrupt(struct mesh_state *ms); ms 188 drivers/scsi/mesh.c static void cmd_complete(struct mesh_state *ms); ms 189 drivers/scsi/mesh.c static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd); ms 190 drivers/scsi/mesh.c static void halt_dma(struct mesh_state *ms); ms 191 drivers/scsi/mesh.c static void phase_mismatch(struct mesh_state *ms); ms 213 drivers/scsi/mesh.c static void dlog(struct mesh_state *ms, char *fmt, int a) ms 215 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 219 drivers/scsi/mesh.c slp = &ms->log[ms->log_ix]; ms 222 drivers/scsi/mesh.c tlp->phase = (ms->msgphase << 4) + ms->phase; ms 223 drivers/scsi/mesh.c tlp->bs0 = ms->mesh->bus_status0; ms 224 drivers/scsi/mesh.c tlp->bs1 = ms->mesh->bus_status1; ms 225 drivers/scsi/mesh.c tlp->tgt = ms->conn_tgt; ms 232 drivers/scsi/mesh.c if (++ms->log_ix >= N_DBG_SLOG) ms 233 drivers/scsi/mesh.c ms->log_ix = 0; ms 234 drivers/scsi/mesh.c if (ms->n_log < N_DBG_SLOG) ms 235 drivers/scsi/mesh.c ++ms->n_log; ms 238 drivers/scsi/mesh.c static void dumplog(struct mesh_state *ms, int t) ms 240 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[t]; ms 264 drivers/scsi/mesh.c static void dumpslog(struct mesh_state *ms) ms 269 drivers/scsi/mesh.c if (ms->n_log == 0) ms 271 drivers/scsi/mesh.c i = ms->log_ix - ms->n_log; ms 274 drivers/scsi/mesh.c ms->n_log = 0; ms 276 drivers/scsi/mesh.c lp = &ms->log[i]; ms 286 drivers/scsi/mesh.c } while (i != ms->log_ix); ms 291 drivers/scsi/mesh.c static inline void dlog(struct mesh_state *ms, char *fmt, int a) ms 293 drivers/scsi/mesh.c static inline void dumplog(struct mesh_state *ms, int tgt) ms 295 drivers/scsi/mesh.c static inline void dumpslog(struct mesh_state *ms) ms 303 drivers/scsi/mesh.c mesh_dump_regs(struct mesh_state *ms) ms 305 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 306 drivers/scsi/mesh.c volatile struct dbdma_regs __iomem *md = ms->dma; ms 311 drivers/scsi/mesh.c ms, mr, md); ms 323 drivers/scsi/mesh.c ms->phase, ms->msgphase, ms->conn_tgt, ms->data_ptr); ms 325 drivers/scsi/mesh.c ms->dma_started, ms->dma_count, ms->n_msgout); ms 327 drivers/scsi/mesh.c tp = &ms->tgts[t]; ms 348 drivers/scsi/mesh.c static void mesh_completed(struct mesh_state *ms, struct scsi_cmnd *cmd) ms 358 drivers/scsi/mesh.c static void mesh_init(struct mesh_state *ms) ms 360 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 361 drivers/scsi/mesh.c volatile struct dbdma_regs __iomem *md = ms->dma; ms 374 drivers/scsi/mesh.c out_8(&mr->source_id, ms->host->this_id); ms 400 drivers/scsi/mesh.c ms->phase = idle; ms 401 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 405 drivers/scsi/mesh.c static void mesh_start_cmd(struct mesh_state *ms, struct scsi_cmnd *cmd) ms 407 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 411 drivers/scsi/mesh.c ms->current_req = cmd; ms 412 drivers/scsi/mesh.c ms->tgts[id].data_goes_out = cmd->sc_data_direction == DMA_TO_DEVICE; ms 413 drivers/scsi/mesh.c ms->tgts[id].current_req = cmd; ms 425 drivers/scsi/mesh.c if (ms->dma_started) ms 428 drivers/scsi/mesh.c ms->phase = arbitrating; ms 429 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 430 drivers/scsi/mesh.c ms->data_ptr = 0; ms 431 drivers/scsi/mesh.c ms->dma_started = 0; ms 432 drivers/scsi/mesh.c ms->n_msgout = 0; ms 433 drivers/scsi/mesh.c ms->last_n_msgout = 0; ms 434 drivers/scsi/mesh.c ms->expect_reply = 0; ms 435 drivers/scsi/mesh.c ms->conn_tgt = id; ms 436 drivers/scsi/mesh.c ms->tgts[id].saved_ptr = 0; ms 437 drivers/scsi/mesh.c ms->stat = DID_OK; ms 438 drivers/scsi/mesh.c ms->aborting = 0; ms 440 drivers/scsi/mesh.c ms->tgts[id].n_log = 0; ms 441 drivers/scsi/mesh.c dlog(ms, "start cmd=%x", (int) cmd); ms 445 drivers/scsi/mesh.c dlog(ms, "about to arb, intr/exc/err/fc=%.8x", ms 457 drivers/scsi/mesh.c dlog(ms, "busy b4 arb, intr/exc/err/fc=%.8x", ms 464 drivers/scsi/mesh.c dlog(ms, "intr b4 arb, intr/exc/err/fc=%.8x", ms 467 drivers/scsi/mesh.c mesh_interrupt(ms); ms 468 drivers/scsi/mesh.c if (ms->phase != arbitrating) ms 475 drivers/scsi/mesh.c ms->stat = DID_BUS_BUSY; ms 476 drivers/scsi/mesh.c ms->phase = idle; ms 477 drivers/scsi/mesh.c mesh_done(ms, 0); ms 502 drivers/scsi/mesh.c dlog(ms, "intr after disresel, intr/exc/err/fc=%.8x", ms 505 drivers/scsi/mesh.c mesh_interrupt(ms); ms 506 drivers/scsi/mesh.c if (ms->phase != arbitrating) ms 508 drivers/scsi/mesh.c dlog(ms, "after intr after disresel, intr/exc/err/fc=%.8x", ms 520 drivers/scsi/mesh.c dlog(ms, "after arb, intr/exc/err/fc=%.8x", ms 525 drivers/scsi/mesh.c dlog(ms, "resel? after arb, intr/exc/err/fc=%.8x", ms 536 drivers/scsi/mesh.c dlog(ms, "tried reset after arb, intr/exc/err/fc=%.8x", ms 558 drivers/scsi/mesh.c static void mesh_start(struct mesh_state *ms) ms 562 drivers/scsi/mesh.c if (ms->phase != idle || ms->current_req != NULL) { ms 564 drivers/scsi/mesh.c ms->phase, ms); ms 568 drivers/scsi/mesh.c while (ms->phase == idle) { ms 570 drivers/scsi/mesh.c for (cmd = ms->request_q; ; cmd = (struct scsi_cmnd *) cmd->host_scribble) { ms 573 drivers/scsi/mesh.c if (ms->tgts[cmd->device->id].current_req == NULL) ms 579 drivers/scsi/mesh.c ms->request_q = next; ms 583 drivers/scsi/mesh.c ms->request_qtail = prev; ms 585 drivers/scsi/mesh.c mesh_start_cmd(ms, cmd); ms 589 drivers/scsi/mesh.c static void mesh_done(struct mesh_state *ms, int start_next) ms 592 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 594 drivers/scsi/mesh.c cmd = ms->current_req; ms 595 drivers/scsi/mesh.c ms->current_req = NULL; ms 598 drivers/scsi/mesh.c cmd->result = (ms->stat << 16) | cmd->SCp.Status; ms 599 drivers/scsi/mesh.c if (ms->stat == DID_OK) ms 603 drivers/scsi/mesh.c cmd->result, ms->data_ptr, scsi_bufflen(cmd)); ms 614 drivers/scsi/mesh.c cmd->SCp.this_residual -= ms->data_ptr; ms 615 drivers/scsi/mesh.c mesh_completed(ms, cmd); ms 618 drivers/scsi/mesh.c out_8(&ms->mesh->sequence, SEQ_ENBRESEL); ms 619 drivers/scsi/mesh.c mesh_flush_io(ms->mesh); ms 621 drivers/scsi/mesh.c ms->phase = idle; ms 622 drivers/scsi/mesh.c mesh_start(ms); ms 626 drivers/scsi/mesh.c static inline void add_sdtr_msg(struct mesh_state *ms) ms 628 drivers/scsi/mesh.c int i = ms->n_msgout; ms 630 drivers/scsi/mesh.c ms->msgout[i] = EXTENDED_MESSAGE; ms 631 drivers/scsi/mesh.c ms->msgout[i+1] = 3; ms 632 drivers/scsi/mesh.c ms->msgout[i+2] = EXTENDED_SDTR; ms 633 drivers/scsi/mesh.c ms->msgout[i+3] = mesh_sync_period/4; ms 634 drivers/scsi/mesh.c ms->msgout[i+4] = (ALLOW_SYNC(ms->conn_tgt)? mesh_sync_offset: 0); ms 635 drivers/scsi/mesh.c ms->n_msgout = i + 5; ms 638 drivers/scsi/mesh.c static void set_sdtr(struct mesh_state *ms, int period, int offset) ms 640 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 641 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 649 drivers/scsi/mesh.c ms->conn_tgt); ms 658 drivers/scsi/mesh.c v = (ms->clk_freq / 5000) * period; ms 663 drivers/scsi/mesh.c tr = (ms->clk_freq + 250000) / 500000; ms 669 drivers/scsi/mesh.c tr = ((ms->clk_freq / (v + 2)) + 199999) / 200000; ms 676 drivers/scsi/mesh.c ms->conn_tgt, tr/10, tr%10); ms 679 drivers/scsi/mesh.c static void start_phase(struct mesh_state *ms) ms 682 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 683 drivers/scsi/mesh.c volatile struct dbdma_regs __iomem *md = ms->dma; ms 684 drivers/scsi/mesh.c struct scsi_cmnd *cmd = ms->current_req; ms 685 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 687 drivers/scsi/mesh.c dlog(ms, "start_phase nmo/exc/fc/seq = %.8x", ms 688 drivers/scsi/mesh.c MKWORD(ms->n_msgout, mr->exception, mr->fifo_count, mr->sequence)); ms 690 drivers/scsi/mesh.c seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0); ms 691 drivers/scsi/mesh.c switch (ms->msgphase) { ms 699 drivers/scsi/mesh.c ms->n_msgin = 0; ms 708 drivers/scsi/mesh.c if (ms->n_msgout <= 0) { ms 710 drivers/scsi/mesh.c ms->n_msgout); ms 711 drivers/scsi/mesh.c mesh_dump_regs(ms); ms 712 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 715 drivers/scsi/mesh.c if (ALLOW_DEBUG(ms->conn_tgt)) { ms 717 drivers/scsi/mesh.c ms->n_msgout); ms 718 drivers/scsi/mesh.c for (i = 0; i < ms->n_msgout; ++i) ms 719 drivers/scsi/mesh.c printk(" %x", ms->msgout[i]); ms 722 drivers/scsi/mesh.c dlog(ms, "msgout msg=%.8x", MKWORD(ms->n_msgout, ms->msgout[0], ms 723 drivers/scsi/mesh.c ms->msgout[1], ms->msgout[2])); ms 733 drivers/scsi/mesh.c dlog(ms, "bus0 was %.2x explicitly asserting ATN", mr->bus_status0); ms 740 drivers/scsi/mesh.c dlog(ms,"hace: after explicit ATN bus0=%.2x",mr->bus_status0); ms 742 drivers/scsi/mesh.c if (ms->n_msgout == 1) { ms 749 drivers/scsi/mesh.c cmd_complete(ms); ms 751 drivers/scsi/mesh.c out_8(&mr->count_lo, ms->n_msgout - 1); ms 753 drivers/scsi/mesh.c for (i = 0; i < ms->n_msgout - 1; ++i) ms 754 drivers/scsi/mesh.c out_8(&mr->fifo, ms->msgout[i]); ms 760 drivers/scsi/mesh.c ms->msgphase); ms 763 drivers/scsi/mesh.c switch (ms->phase) { ms 765 drivers/scsi/mesh.c out_8(&mr->dest_id, ms->conn_tgt); ms 785 drivers/scsi/mesh.c if (!ms->dma_started) { ms 786 drivers/scsi/mesh.c set_dma_cmds(ms, cmd); ms 787 drivers/scsi/mesh.c out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds)); ms 789 drivers/scsi/mesh.c ms->dma_started = 1; ms 791 drivers/scsi/mesh.c nb = ms->dma_count; ms 794 drivers/scsi/mesh.c ms->dma_count -= nb; ms 795 drivers/scsi/mesh.c ms->data_ptr += nb; ms 811 drivers/scsi/mesh.c dlog(ms, "enbresel intr/exc/err/fc=%.8x", ms 818 drivers/scsi/mesh.c ms->phase); ms 819 drivers/scsi/mesh.c dumpslog(ms); ms 824 drivers/scsi/mesh.c static inline void get_msgin(struct mesh_state *ms) ms 826 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 831 drivers/scsi/mesh.c i = ms->n_msgin; ms 832 drivers/scsi/mesh.c ms->n_msgin = i + n; ms 834 drivers/scsi/mesh.c ms->msgin[i++] = in_8(&mr->fifo); ms 838 drivers/scsi/mesh.c static inline int msgin_length(struct mesh_state *ms) ms 843 drivers/scsi/mesh.c if (ms->n_msgin > 0) { ms 844 drivers/scsi/mesh.c b = ms->msgin[0]; ms 847 drivers/scsi/mesh.c n = ms->n_msgin < 2? 2: ms->msgin[1] + 2; ms 856 drivers/scsi/mesh.c static void reselected(struct mesh_state *ms) ms 858 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 863 drivers/scsi/mesh.c switch (ms->phase) { ms 867 drivers/scsi/mesh.c if ((cmd = ms->current_req) != NULL) { ms 869 drivers/scsi/mesh.c cmd->host_scribble = (void *) ms->request_q; ms 870 drivers/scsi/mesh.c if (ms->request_q == NULL) ms 871 drivers/scsi/mesh.c ms->request_qtail = cmd; ms 872 drivers/scsi/mesh.c ms->request_q = cmd; ms 873 drivers/scsi/mesh.c tp = &ms->tgts[cmd->device->id]; ms 878 drivers/scsi/mesh.c ms->phase = reselecting; ms 879 drivers/scsi/mesh.c mesh_done(ms, 0); ms 885 drivers/scsi/mesh.c ms->msgphase, ms->phase, ms->conn_tgt); ms 886 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 887 drivers/scsi/mesh.c dumpslog(ms); ms 890 drivers/scsi/mesh.c if (ms->dma_started) { ms 892 drivers/scsi/mesh.c halt_dma(ms); ms 894 drivers/scsi/mesh.c ms->current_req = NULL; ms 895 drivers/scsi/mesh.c ms->phase = dataing; ms 896 drivers/scsi/mesh.c ms->msgphase = msg_in; ms 897 drivers/scsi/mesh.c ms->n_msgout = 0; ms 898 drivers/scsi/mesh.c ms->last_n_msgout = 0; ms 899 drivers/scsi/mesh.c prev = ms->conn_tgt; ms 913 drivers/scsi/mesh.c dlog(ms, "extra resel err/exc/fc = %.6x", ms 929 drivers/scsi/mesh.c ms->conn_tgt = ms->host->this_id; ms 935 drivers/scsi/mesh.c dlog(ms, "reseldata %x", b); ms 938 drivers/scsi/mesh.c if ((b & (1 << t)) != 0 && t != ms->host->this_id) ms 940 drivers/scsi/mesh.c if (b != (1 << t) + (1 << ms->host->this_id)) { ms 942 drivers/scsi/mesh.c ms->conn_tgt = ms->host->this_id; ms 950 drivers/scsi/mesh.c ms->conn_tgt = t; ms 951 drivers/scsi/mesh.c tp = &ms->tgts[t]; ms 958 drivers/scsi/mesh.c ms->current_req = tp->current_req; ms 963 drivers/scsi/mesh.c ms->data_ptr = tp->saved_ptr; ms 964 drivers/scsi/mesh.c dlog(ms, "resel prev tgt=%d", prev); ms 965 drivers/scsi/mesh.c dlog(ms, "resel err/exc=%.4x", MKWORD(0, 0, mr->error, mr->exception)); ms 966 drivers/scsi/mesh.c start_phase(ms); ms 970 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 971 drivers/scsi/mesh.c dumpslog(ms); ms 972 drivers/scsi/mesh.c ms->data_ptr = 0; ms 973 drivers/scsi/mesh.c ms->aborting = 1; ms 974 drivers/scsi/mesh.c start_phase(ms); ms 977 drivers/scsi/mesh.c static void do_abort(struct mesh_state *ms) ms 979 drivers/scsi/mesh.c ms->msgout[0] = ABORT; ms 980 drivers/scsi/mesh.c ms->n_msgout = 1; ms 981 drivers/scsi/mesh.c ms->aborting = 1; ms 982 drivers/scsi/mesh.c ms->stat = DID_ABORT; ms 983 drivers/scsi/mesh.c dlog(ms, "abort", 0); ms 986 drivers/scsi/mesh.c static void handle_reset(struct mesh_state *ms) ms 991 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 994 drivers/scsi/mesh.c tp = &ms->tgts[tgt]; ms 998 drivers/scsi/mesh.c mesh_completed(ms, cmd); ms 1000 drivers/scsi/mesh.c ms->tgts[tgt].sdtr_state = do_sdtr; ms 1001 drivers/scsi/mesh.c ms->tgts[tgt].sync_params = ASYNC_PARAMS; ms 1003 drivers/scsi/mesh.c ms->current_req = NULL; ms 1004 drivers/scsi/mesh.c while ((cmd = ms->request_q) != NULL) { ms 1005 drivers/scsi/mesh.c ms->request_q = (struct scsi_cmnd *) cmd->host_scribble; ms 1007 drivers/scsi/mesh.c mesh_completed(ms, cmd); ms 1009 drivers/scsi/mesh.c ms->phase = idle; ms 1010 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 1022 drivers/scsi/mesh.c struct mesh_state *ms = dev_id; ms 1023 drivers/scsi/mesh.c struct Scsi_Host *dev = ms->host; ms 1026 drivers/scsi/mesh.c mesh_interrupt(ms); ms 1031 drivers/scsi/mesh.c static void handle_error(struct mesh_state *ms) ms 1034 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1039 drivers/scsi/mesh.c dlog(ms, "error err/exc/fc/cl=%.8x", ms 1048 drivers/scsi/mesh.c handle_reset(ms); ms 1055 drivers/scsi/mesh.c reselected(ms); ms 1058 drivers/scsi/mesh.c if (!ms->aborting) { ms 1060 drivers/scsi/mesh.c ms->conn_tgt); ms 1061 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 1062 drivers/scsi/mesh.c dumpslog(ms); ms 1065 drivers/scsi/mesh.c ms->stat = DID_ABORT; ms 1066 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1070 drivers/scsi/mesh.c if (ms->msgphase == msg_in) { ms 1072 drivers/scsi/mesh.c ms->conn_tgt); ms 1073 drivers/scsi/mesh.c ms->msgout[0] = MSG_PARITY_ERROR; ms 1074 drivers/scsi/mesh.c ms->n_msgout = 1; ms 1075 drivers/scsi/mesh.c ms->msgphase = msg_in_bad; ms 1076 drivers/scsi/mesh.c cmd_complete(ms); ms 1079 drivers/scsi/mesh.c if (ms->stat == DID_OK) { ms 1081 drivers/scsi/mesh.c ms->conn_tgt); ms 1082 drivers/scsi/mesh.c ms->stat = DID_PARITY; ms 1086 drivers/scsi/mesh.c cmd_complete(ms); ms 1099 drivers/scsi/mesh.c reselected(ms); ms 1105 drivers/scsi/mesh.c phase_mismatch(ms); ms 1113 drivers/scsi/mesh.c mesh_dump_regs(ms); ms 1114 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 1115 drivers/scsi/mesh.c if (ms->phase > selecting && (in_8(&mr->bus_status1) & BS1_BSY)) { ms 1117 drivers/scsi/mesh.c do_abort(ms); ms 1118 drivers/scsi/mesh.c phase_mismatch(ms); ms 1121 drivers/scsi/mesh.c ms->stat = DID_ERROR; ms 1122 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1125 drivers/scsi/mesh.c static void handle_exception(struct mesh_state *ms) ms 1128 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1135 drivers/scsi/mesh.c reselected(ms); ms 1138 drivers/scsi/mesh.c ms->stat = DID_BUS_BUSY; ms 1139 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1142 drivers/scsi/mesh.c ms->stat = DID_BAD_TARGET; ms 1143 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1147 drivers/scsi/mesh.c phase_mismatch(ms); ms 1150 drivers/scsi/mesh.c mesh_dump_regs(ms); ms 1151 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 1152 drivers/scsi/mesh.c do_abort(ms); ms 1153 drivers/scsi/mesh.c phase_mismatch(ms); ms 1157 drivers/scsi/mesh.c static void handle_msgin(struct mesh_state *ms) ms 1160 drivers/scsi/mesh.c struct scsi_cmnd *cmd = ms->current_req; ms 1161 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 1163 drivers/scsi/mesh.c if (ms->n_msgin == 0) ms 1165 drivers/scsi/mesh.c code = ms->msgin[0]; ms 1166 drivers/scsi/mesh.c if (ALLOW_DEBUG(ms->conn_tgt)) { ms 1167 drivers/scsi/mesh.c printk(KERN_DEBUG "got %d message bytes:", ms->n_msgin); ms 1168 drivers/scsi/mesh.c for (i = 0; i < ms->n_msgin; ++i) ms 1169 drivers/scsi/mesh.c printk(" %x", ms->msgin[i]); ms 1172 drivers/scsi/mesh.c dlog(ms, "msgin msg=%.8x", ms 1173 drivers/scsi/mesh.c MKWORD(ms->n_msgin, code, ms->msgin[1], ms->msgin[2])); ms 1175 drivers/scsi/mesh.c ms->expect_reply = 0; ms 1176 drivers/scsi/mesh.c ms->n_msgout = 0; ms 1177 drivers/scsi/mesh.c if (ms->n_msgin < msgin_length(ms)) ms 1185 drivers/scsi/mesh.c switch (ms->msgin[2]) { ms 1187 drivers/scsi/mesh.c ms->data_ptr += (ms->msgin[3] << 24) + ms->msgin[6] ms 1188 drivers/scsi/mesh.c + (ms->msgin[4] << 16) + (ms->msgin[5] << 8); ms 1193 drivers/scsi/mesh.c add_sdtr_msg(ms); ms 1196 drivers/scsi/mesh.c if (ms->msgout[3] < ms->msgin[3]) ms 1197 drivers/scsi/mesh.c ms->msgout[3] = ms->msgin[3]; ms 1198 drivers/scsi/mesh.c if (ms->msgout[4] > ms->msgin[4]) ms 1199 drivers/scsi/mesh.c ms->msgout[4] = ms->msgin[4]; ms 1200 drivers/scsi/mesh.c set_sdtr(ms, ms->msgout[3], ms->msgout[4]); ms 1201 drivers/scsi/mesh.c ms->msgphase = msg_out; ms 1203 drivers/scsi/mesh.c set_sdtr(ms, ms->msgin[3], ms->msgin[4]); ms 1211 drivers/scsi/mesh.c tp->saved_ptr = ms->data_ptr; ms 1214 drivers/scsi/mesh.c ms->data_ptr = tp->saved_ptr; ms 1217 drivers/scsi/mesh.c ms->phase = disconnecting; ms 1223 drivers/scsi/mesh.c set_sdtr(ms, 0, 0); ms 1230 drivers/scsi/mesh.c do_abort(ms); ms 1231 drivers/scsi/mesh.c ms->msgphase = msg_out; ms 1236 drivers/scsi/mesh.c cmd->device->lun, ms->conn_tgt); ms 1246 drivers/scsi/mesh.c ms->conn_tgt); ms 1247 drivers/scsi/mesh.c for (i = 0; i < ms->n_msgin; ++i) ms 1248 drivers/scsi/mesh.c printk(" %x", ms->msgin[i]); ms 1250 drivers/scsi/mesh.c ms->msgout[0] = MESSAGE_REJECT; ms 1251 drivers/scsi/mesh.c ms->n_msgout = 1; ms 1252 drivers/scsi/mesh.c ms->msgphase = msg_out; ms 1258 drivers/scsi/mesh.c static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd) ms 1264 drivers/scsi/mesh.c dma_cmd = ms->tgts[ms->conn_tgt].data_goes_out? ms 1266 drivers/scsi/mesh.c dcmds = ms->dma_cmds; ms 1278 drivers/scsi/mesh.c off = ms->data_ptr; ms 1316 drivers/scsi/mesh.c ms->dma_count = dtot; ms 1319 drivers/scsi/mesh.c static void halt_dma(struct mesh_state *ms) ms 1321 drivers/scsi/mesh.c volatile struct dbdma_regs __iomem *md = ms->dma; ms 1322 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1323 drivers/scsi/mesh.c struct scsi_cmnd *cmd = ms->current_req; ms 1326 drivers/scsi/mesh.c if (!ms->tgts[ms->conn_tgt].data_goes_out) { ms 1337 drivers/scsi/mesh.c dlog(ms, "halt_dma fc/count=%.6x", ms 1339 drivers/scsi/mesh.c if (ms->tgts[ms->conn_tgt].data_goes_out) ms 1343 drivers/scsi/mesh.c ms->data_ptr -= nb; ms 1344 drivers/scsi/mesh.c dlog(ms, "data_ptr %x", ms->data_ptr); ms 1345 drivers/scsi/mesh.c if (ms->data_ptr < 0) { ms 1347 drivers/scsi/mesh.c ms->data_ptr, nb, ms); ms 1348 drivers/scsi/mesh.c ms->data_ptr = 0; ms 1350 drivers/scsi/mesh.c dumplog(ms, ms->conn_tgt); ms 1351 drivers/scsi/mesh.c dumpslog(ms); ms 1354 drivers/scsi/mesh.c ms->data_ptr > scsi_bufflen(cmd)) { ms 1357 drivers/scsi/mesh.c ms->conn_tgt, ms->data_ptr, scsi_bufflen(cmd), ms 1358 drivers/scsi/mesh.c ms->tgts[ms->conn_tgt].data_goes_out); ms 1361 drivers/scsi/mesh.c ms->dma_started = 0; ms 1364 drivers/scsi/mesh.c static void phase_mismatch(struct mesh_state *ms) ms 1366 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1369 drivers/scsi/mesh.c dlog(ms, "phasemm ch/cl/seq/fc=%.8x", ms 1372 drivers/scsi/mesh.c if (ms->msgphase == msg_out_xxx && phase == BP_MSGOUT) { ms 1378 drivers/scsi/mesh.c out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]); ms 1379 drivers/scsi/mesh.c ms->msgphase = msg_out_last; ms 1383 drivers/scsi/mesh.c if (ms->msgphase == msg_in) { ms 1384 drivers/scsi/mesh.c get_msgin(ms); ms 1385 drivers/scsi/mesh.c if (ms->n_msgin) ms 1386 drivers/scsi/mesh.c handle_msgin(ms); ms 1389 drivers/scsi/mesh.c if (ms->dma_started) ms 1390 drivers/scsi/mesh.c halt_dma(ms); ms 1397 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 1400 drivers/scsi/mesh.c ms->tgts[ms->conn_tgt].data_goes_out = 0; ms 1401 drivers/scsi/mesh.c ms->phase = dataing; ms 1404 drivers/scsi/mesh.c ms->tgts[ms->conn_tgt].data_goes_out = 1; ms 1405 drivers/scsi/mesh.c ms->phase = dataing; ms 1408 drivers/scsi/mesh.c ms->phase = commanding; ms 1411 drivers/scsi/mesh.c ms->phase = statusing; ms 1414 drivers/scsi/mesh.c ms->msgphase = msg_in; ms 1415 drivers/scsi/mesh.c ms->n_msgin = 0; ms 1418 drivers/scsi/mesh.c ms->msgphase = msg_out; ms 1419 drivers/scsi/mesh.c if (ms->n_msgout == 0) { ms 1420 drivers/scsi/mesh.c if (ms->aborting) { ms 1421 drivers/scsi/mesh.c do_abort(ms); ms 1423 drivers/scsi/mesh.c if (ms->last_n_msgout == 0) { ms 1426 drivers/scsi/mesh.c ms->msgout[0] = NOP; ms 1427 drivers/scsi/mesh.c ms->last_n_msgout = 1; ms 1429 drivers/scsi/mesh.c ms->n_msgout = ms->last_n_msgout; ms 1435 drivers/scsi/mesh.c ms->stat = DID_ERROR; ms 1436 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1440 drivers/scsi/mesh.c start_phase(ms); ms 1443 drivers/scsi/mesh.c static void cmd_complete(struct mesh_state *ms) ms 1445 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1446 drivers/scsi/mesh.c struct scsi_cmnd *cmd = ms->current_req; ms 1447 drivers/scsi/mesh.c struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; ms 1450 drivers/scsi/mesh.c dlog(ms, "cmd_complete fc=%x", mr->fifo_count); ms 1451 drivers/scsi/mesh.c seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0); ms 1452 drivers/scsi/mesh.c switch (ms->msgphase) { ms 1455 drivers/scsi/mesh.c ms->n_msgin = 0; ms 1456 drivers/scsi/mesh.c ms->msgphase = msg_in; ms 1461 drivers/scsi/mesh.c get_msgin(ms); ms 1462 drivers/scsi/mesh.c n = msgin_length(ms); ms 1463 drivers/scsi/mesh.c if (ms->n_msgin < n) { ms 1464 drivers/scsi/mesh.c out_8(&mr->count_lo, n - ms->n_msgin); ms 1467 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 1468 drivers/scsi/mesh.c handle_msgin(ms); ms 1469 drivers/scsi/mesh.c start_phase(ms); ms 1498 drivers/scsi/mesh.c dlog(ms, "last_mbyte err/exc/fc/cl=%.8x", ms 1503 drivers/scsi/mesh.c ms->last_n_msgout = ms->n_msgout; ms 1504 drivers/scsi/mesh.c ms->n_msgout = 0; ms 1508 drivers/scsi/mesh.c handle_error(ms); ms 1517 drivers/scsi/mesh.c handle_exception(ms); ms 1524 drivers/scsi/mesh.c out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]); ms 1525 drivers/scsi/mesh.c ms->msgphase = msg_out_last; ms 1528 drivers/scsi/mesh.c ms->msgphase = msg_out_xxx; ms 1533 drivers/scsi/mesh.c ms->last_n_msgout = ms->n_msgout; ms 1534 drivers/scsi/mesh.c ms->n_msgout = 0; ms 1535 drivers/scsi/mesh.c ms->msgphase = ms->expect_reply? msg_in: msg_none; ms 1536 drivers/scsi/mesh.c start_phase(ms); ms 1540 drivers/scsi/mesh.c switch (ms->phase) { ms 1543 drivers/scsi/mesh.c dumpslog(ms); ms 1546 drivers/scsi/mesh.c dlog(ms, "Selecting phase at command completion",0); ms 1547 drivers/scsi/mesh.c ms->msgout[0] = IDENTIFY(ALLOW_RESEL(ms->conn_tgt), ms 1549 drivers/scsi/mesh.c ms->n_msgout = 1; ms 1550 drivers/scsi/mesh.c ms->expect_reply = 0; ms 1551 drivers/scsi/mesh.c if (ms->aborting) { ms 1552 drivers/scsi/mesh.c ms->msgout[0] = ABORT; ms 1553 drivers/scsi/mesh.c ms->n_msgout++; ms 1556 drivers/scsi/mesh.c add_sdtr_msg(ms); ms 1557 drivers/scsi/mesh.c ms->expect_reply = 1; ms 1560 drivers/scsi/mesh.c ms->msgphase = msg_out; ms 1571 drivers/scsi/mesh.c dlog(ms, "impatient for req", ms->n_msgout); ms 1572 drivers/scsi/mesh.c ms->msgphase = msg_none; ms 1579 drivers/scsi/mesh.c if (ms->dma_count != 0) { ms 1580 drivers/scsi/mesh.c start_phase(ms); ms 1596 drivers/scsi/mesh.c halt_dma(ms); ms 1605 drivers/scsi/mesh.c ms->msgphase = msg_in; ms 1608 drivers/scsi/mesh.c mesh_done(ms, 1); ms 1611 drivers/scsi/mesh.c ms->current_req = NULL; ms 1612 drivers/scsi/mesh.c ms->phase = idle; ms 1613 drivers/scsi/mesh.c mesh_start(ms); ms 1618 drivers/scsi/mesh.c ++ms->phase; ms 1619 drivers/scsi/mesh.c start_phase(ms); ms 1631 drivers/scsi/mesh.c struct mesh_state *ms; ms 1636 drivers/scsi/mesh.c ms = (struct mesh_state *) cmd->device->host->hostdata; ms 1638 drivers/scsi/mesh.c if (ms->request_q == NULL) ms 1639 drivers/scsi/mesh.c ms->request_q = cmd; ms 1641 drivers/scsi/mesh.c ms->request_qtail->host_scribble = (void *) cmd; ms 1642 drivers/scsi/mesh.c ms->request_qtail = cmd; ms 1644 drivers/scsi/mesh.c if (ms->phase == idle) ms 1645 drivers/scsi/mesh.c mesh_start(ms); ms 1657 drivers/scsi/mesh.c static void mesh_interrupt(struct mesh_state *ms) ms 1659 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1663 drivers/scsi/mesh.c if (ALLOW_DEBUG(ms->conn_tgt)) ms 1667 drivers/scsi/mesh.c ms->phase, ms->msgphase); ms 1670 drivers/scsi/mesh.c dlog(ms, "interrupt intr/err/exc/seq=%.8x", ms 1673 drivers/scsi/mesh.c handle_error(ms); ms 1675 drivers/scsi/mesh.c handle_exception(ms); ms 1678 drivers/scsi/mesh.c cmd_complete(ms); ms 1689 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata; ms 1692 drivers/scsi/mesh.c mesh_dump_regs(ms); ms 1693 drivers/scsi/mesh.c dumplog(ms, cmd->device->id); ms 1694 drivers/scsi/mesh.c dumpslog(ms); ms 1706 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata; ms 1707 drivers/scsi/mesh.c volatile struct mesh_regs __iomem *mr = ms->mesh; ms 1708 drivers/scsi/mesh.c volatile struct dbdma_regs __iomem *md = ms->dma; ms 1713 drivers/scsi/mesh.c spin_lock_irqsave(ms->host->host_lock, flags); ms 1723 drivers/scsi/mesh.c out_8(&mr->source_id, ms->host->this_id); ms 1734 drivers/scsi/mesh.c handle_reset(ms); ms 1736 drivers/scsi/mesh.c spin_unlock_irqrestore(ms->host->host_lock, flags); ms 1740 drivers/scsi/mesh.c static void set_mesh_power(struct mesh_state *ms, int state) ms 1745 drivers/scsi/mesh.c pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 1); ms 1748 drivers/scsi/mesh.c pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 0); ms 1757 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); ms 1768 drivers/scsi/mesh.c if (ms->phase == sleeping) ms 1771 drivers/scsi/mesh.c scsi_block_requests(ms->host); ms 1772 drivers/scsi/mesh.c spin_lock_irqsave(ms->host->host_lock, flags); ms 1773 drivers/scsi/mesh.c while(ms->phase != idle) { ms 1774 drivers/scsi/mesh.c spin_unlock_irqrestore(ms->host->host_lock, flags); ms 1776 drivers/scsi/mesh.c spin_lock_irqsave(ms->host->host_lock, flags); ms 1778 drivers/scsi/mesh.c ms->phase = sleeping; ms 1779 drivers/scsi/mesh.c spin_unlock_irqrestore(ms->host->host_lock, flags); ms 1780 drivers/scsi/mesh.c disable_irq(ms->meshintr); ms 1781 drivers/scsi/mesh.c set_mesh_power(ms, 0); ms 1788 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); ms 1791 drivers/scsi/mesh.c if (ms->phase != sleeping) ms 1794 drivers/scsi/mesh.c set_mesh_power(ms, 1); ms 1795 drivers/scsi/mesh.c mesh_init(ms); ms 1796 drivers/scsi/mesh.c spin_lock_irqsave(ms->host->host_lock, flags); ms 1797 drivers/scsi/mesh.c mesh_start(ms); ms 1798 drivers/scsi/mesh.c spin_unlock_irqrestore(ms->host->host_lock, flags); ms 1799 drivers/scsi/mesh.c enable_irq(ms->meshintr); ms 1800 drivers/scsi/mesh.c scsi_unblock_requests(ms->host); ms 1814 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); ms 1819 drivers/scsi/mesh.c spin_lock_irqsave(ms->host->host_lock, flags); ms 1820 drivers/scsi/mesh.c mr = ms->mesh; ms 1827 drivers/scsi/mesh.c spin_unlock_irqrestore(ms->host->host_lock, flags); ms 1851 drivers/scsi/mesh.c struct mesh_state *ms; ms 1890 drivers/scsi/mesh.c ms = (struct mesh_state *) mesh_host->hostdata; ms 1891 drivers/scsi/mesh.c macio_set_drvdata(mdev, ms); ms 1892 drivers/scsi/mesh.c ms->host = mesh_host; ms 1893 drivers/scsi/mesh.c ms->mdev = mdev; ms 1894 drivers/scsi/mesh.c ms->pdev = pdev; ms 1896 drivers/scsi/mesh.c ms->mesh = ioremap(macio_resource_start(mdev, 0), 0x1000); ms 1897 drivers/scsi/mesh.c if (ms->mesh == NULL) { ms 1901 drivers/scsi/mesh.c ms->dma = ioremap(macio_resource_start(mdev, 1), 0x1000); ms 1902 drivers/scsi/mesh.c if (ms->dma == NULL) { ms 1904 drivers/scsi/mesh.c iounmap(ms->mesh); ms 1908 drivers/scsi/mesh.c ms->meshintr = macio_irq(mdev, 0); ms 1909 drivers/scsi/mesh.c ms->dmaintr = macio_irq(mdev, 1); ms 1914 drivers/scsi/mesh.c ms->dma_cmd_size = (mesh_host->sg_tablesize + 2) * sizeof(struct dbdma_cmd); ms 1920 drivers/scsi/mesh.c ms->dma_cmd_size, &dma_cmd_bus, ms 1927 drivers/scsi/mesh.c ms->dma_cmds = (struct dbdma_cmd *) DBDMA_ALIGN(dma_cmd_space); ms 1928 drivers/scsi/mesh.c ms->dma_cmd_space = dma_cmd_space; ms 1929 drivers/scsi/mesh.c ms->dma_cmd_bus = dma_cmd_bus + ((unsigned long)ms->dma_cmds) ms 1931 drivers/scsi/mesh.c ms->current_req = NULL; ms 1933 drivers/scsi/mesh.c ms->tgts[tgt].sdtr_state = do_sdtr; ms 1934 drivers/scsi/mesh.c ms->tgts[tgt].sync_params = ASYNC_PARAMS; ms 1935 drivers/scsi/mesh.c ms->tgts[tgt].current_req = NULL; ms 1939 drivers/scsi/mesh.c ms->clk_freq = *cfp; ms 1942 drivers/scsi/mesh.c ms->clk_freq = 50000000; ms 1948 drivers/scsi/mesh.c minper = 1000000000 / (ms->clk_freq / 5); /* ns */ ms 1953 drivers/scsi/mesh.c set_mesh_power(ms, 1); ms 1956 drivers/scsi/mesh.c mesh_init(ms); ms 1959 drivers/scsi/mesh.c if (request_irq(ms->meshintr, do_mesh_interrupt, 0, "MESH", ms)) { ms 1960 drivers/scsi/mesh.c printk(KERN_ERR "MESH: can't get irq %d\n", ms->meshintr); ms 1972 drivers/scsi/mesh.c free_irq(ms->meshintr, ms); ms 1978 drivers/scsi/mesh.c set_mesh_power(ms, 0); ms 1979 drivers/scsi/mesh.c dma_free_coherent(&macio_get_pci_dev(mdev)->dev, ms->dma_cmd_size, ms 1980 drivers/scsi/mesh.c ms->dma_cmd_space, ms->dma_cmd_bus); ms 1982 drivers/scsi/mesh.c iounmap(ms->dma); ms 1983 drivers/scsi/mesh.c iounmap(ms->mesh); ms 1994 drivers/scsi/mesh.c struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); ms 1995 drivers/scsi/mesh.c struct Scsi_Host *mesh_host = ms->host; ms 1999 drivers/scsi/mesh.c free_irq(ms->meshintr, ms); ms 2005 drivers/scsi/mesh.c set_mesh_power(ms, 0); ms 2008 drivers/scsi/mesh.c iounmap(ms->mesh); ms 2009 drivers/scsi/mesh.c iounmap(ms->dma); ms 2012 drivers/scsi/mesh.c dma_free_coherent(&macio_get_pci_dev(mdev)->dev, ms->dma_cmd_size, ms 2013 drivers/scsi/mesh.c ms->dma_cmd_space, ms->dma_cmd_bus); ms 7843 drivers/scsi/ncr53c8xx.c unsigned ms = 0; ms 7869 drivers/scsi/ncr53c8xx.c while (!(INW(nc_sist) & GEN) && ms++ < 100000) { ms 7882 drivers/scsi/ncr53c8xx.c printk ("%s: Delay (GEN=%d): %u msec\n", ncr_name(np), gen, ms); ms 7886 drivers/scsi/ncr53c8xx.c return ms ? ((1 << gen) * 4340) / ms : 0; ms 872 drivers/scsi/sg.c unsigned int ms; ms 887 drivers/scsi/sg.c ms = jiffies_to_msecs(jiffies); ms 889 drivers/scsi/sg.c (ms > srp->header.duration) ? ms 890 drivers/scsi/sg.c (ms - srp->header.duration) : 0; ms 1327 drivers/scsi/sg.c unsigned int ms; ms 1350 drivers/scsi/sg.c ms = jiffies_to_msecs(jiffies); ms 1351 drivers/scsi/sg.c srp->header.duration = (ms > srp->header.duration) ? ms 1352 drivers/scsi/sg.c (ms - srp->header.duration) : 0; ms 2524 drivers/scsi/sg.c unsigned int ms; ms 2565 drivers/scsi/sg.c ms = jiffies_to_msecs(jiffies); ms 2569 drivers/scsi/sg.c (ms > hp->duration ? ms - hp->duration : 0)); ms 291 drivers/scsi/sym53c8xx_2/sym_hipd.c unsigned int ms = 0; ms 325 drivers/scsi/sym53c8xx_2/sym_hipd.c while (!(INW(np, nc_sist) & GEN) && ms++ < 100000) ms 345 drivers/scsi/sym53c8xx_2/sym_hipd.c f = ms ? ((1 << gen) * (4340*4)) / ms : 0; ms 356 drivers/scsi/sym53c8xx_2/sym_hipd.c sym_name(np), gen, ms/4, f); ms 151 drivers/slimbus/messaging.c unsigned long ms = txn->rl + HZ; ms 154 drivers/slimbus/messaging.c msecs_to_jiffies(ms)); ms 331 drivers/slimbus/qcom-ctrl.c unsigned long ms = txn->rl + HZ; ms 377 drivers/slimbus/qcom-ctrl.c timeout = wait_for_completion_timeout(&done, msecs_to_jiffies(ms)); ms 82 drivers/spi/spi-mpc52xx.c int (*state)(int irq, struct mpc52xx_spi *ms, u8 status, u8 data); ms 95 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_chipsel(struct mpc52xx_spi *ms, int value) ms 99 drivers/spi/spi-mpc52xx.c if (ms->gpio_cs_count > 0) { ms 100 drivers/spi/spi-mpc52xx.c cs = ms->message->spi->chip_select; ms 101 drivers/spi/spi-mpc52xx.c gpio_set_value(ms->gpio_cs[cs], value ? 0 : 1); ms 103 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_PORTDATA, value ? 0 : 0x08); ms 111 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_start_transfer(struct mpc52xx_spi *ms) ms 113 drivers/spi/spi-mpc52xx.c ms->rx_buf = ms->transfer->rx_buf; ms 114 drivers/spi/spi-mpc52xx.c ms->tx_buf = ms->transfer->tx_buf; ms 115 drivers/spi/spi-mpc52xx.c ms->len = ms->transfer->len; ms 118 drivers/spi/spi-mpc52xx.c if (ms->cs_change) ms 119 drivers/spi/spi-mpc52xx.c mpc52xx_spi_chipsel(ms, 1); ms 120 drivers/spi/spi-mpc52xx.c ms->cs_change = ms->transfer->cs_change; ms 123 drivers/spi/spi-mpc52xx.c ms->wcol_tx_timestamp = get_tbl(); ms 124 drivers/spi/spi-mpc52xx.c if (ms->tx_buf) ms 125 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_DATA, *ms->tx_buf++); ms 127 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_DATA, 0); ms 131 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms, ms 133 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms, ms 143 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsmstate_idle(int irq, struct mpc52xx_spi *ms, u8 status, u8 data) ms 150 drivers/spi/spi-mpc52xx.c dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n", ms 154 drivers/spi/spi-mpc52xx.c if (list_empty(&ms->queue)) ms 158 drivers/spi/spi-mpc52xx.c ms->message = list_first_entry(&ms->queue, struct spi_message, queue); ms 159 drivers/spi/spi-mpc52xx.c list_del_init(&ms->message->queue); ms 163 drivers/spi/spi-mpc52xx.c spi = ms->message->spi; ms 170 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_CTRL1, ctrl1); ms 175 drivers/spi/spi-mpc52xx.c sppr = ((ms->ipb_freq / ms->message->spi->max_speed_hz) + 1) >> 1; ms 189 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_BRR, sppr << 4 | spr); /* Set speed */ ms 191 drivers/spi/spi-mpc52xx.c ms->cs_change = 1; ms 192 drivers/spi/spi-mpc52xx.c ms->transfer = container_of(ms->message->transfers.next, ms 195 drivers/spi/spi-mpc52xx.c mpc52xx_spi_start_transfer(ms); ms 196 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_transfer; ms 209 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms, ms 213 drivers/spi/spi-mpc52xx.c return ms->irq0 ? FSM_STOP : FSM_POLL; ms 223 drivers/spi/spi-mpc52xx.c ms->wcol_count++; ms 224 drivers/spi/spi-mpc52xx.c ms->wcol_ticks += get_tbl() - ms->wcol_tx_timestamp; ms 225 drivers/spi/spi-mpc52xx.c ms->wcol_tx_timestamp = get_tbl(); ms 227 drivers/spi/spi-mpc52xx.c if (ms->tx_buf) ms 228 drivers/spi/spi-mpc52xx.c data = *(ms->tx_buf - 1); ms 229 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_DATA, data); /* try again */ ms 232 drivers/spi/spi-mpc52xx.c ms->modf_count++; ms 233 drivers/spi/spi-mpc52xx.c dev_err(&ms->master->dev, "mode fault\n"); ms 234 drivers/spi/spi-mpc52xx.c mpc52xx_spi_chipsel(ms, 0); ms 235 drivers/spi/spi-mpc52xx.c ms->message->status = -EIO; ms 236 drivers/spi/spi-mpc52xx.c if (ms->message->complete) ms 237 drivers/spi/spi-mpc52xx.c ms->message->complete(ms->message->context); ms 238 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_idle; ms 243 drivers/spi/spi-mpc52xx.c ms->byte_count++; ms 244 drivers/spi/spi-mpc52xx.c if (ms->rx_buf) ms 245 drivers/spi/spi-mpc52xx.c *ms->rx_buf++ = data; ms 248 drivers/spi/spi-mpc52xx.c ms->len--; ms 249 drivers/spi/spi-mpc52xx.c if (ms->len == 0) { ms 250 drivers/spi/spi-mpc52xx.c ms->timestamp = get_tbl(); ms 251 drivers/spi/spi-mpc52xx.c ms->timestamp += ms->transfer->delay_usecs * tb_ticks_per_usec; ms 252 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_wait; ms 257 drivers/spi/spi-mpc52xx.c ms->wcol_tx_timestamp = get_tbl(); ms 258 drivers/spi/spi-mpc52xx.c if (ms->tx_buf) ms 259 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_DATA, *ms->tx_buf++); ms 261 drivers/spi/spi-mpc52xx.c out_8(ms->regs + SPI_DATA, 0); ms 273 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms, u8 status, u8 data) ms 276 drivers/spi/spi-mpc52xx.c dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n", ms 279 drivers/spi/spi-mpc52xx.c if (((int)get_tbl()) - ms->timestamp < 0) ms 282 drivers/spi/spi-mpc52xx.c ms->message->actual_length += ms->transfer->len; ms 287 drivers/spi/spi-mpc52xx.c if (ms->transfer->transfer_list.next == &ms->message->transfers) { ms 288 drivers/spi/spi-mpc52xx.c ms->msg_count++; ms 289 drivers/spi/spi-mpc52xx.c mpc52xx_spi_chipsel(ms, 0); ms 290 drivers/spi/spi-mpc52xx.c ms->message->status = 0; ms 291 drivers/spi/spi-mpc52xx.c if (ms->message->complete) ms 292 drivers/spi/spi-mpc52xx.c ms->message->complete(ms->message->context); ms 293 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_idle; ms 299 drivers/spi/spi-mpc52xx.c if (ms->cs_change) ms 300 drivers/spi/spi-mpc52xx.c mpc52xx_spi_chipsel(ms, 0); ms 302 drivers/spi/spi-mpc52xx.c ms->transfer = container_of(ms->transfer->transfer_list.next, ms 304 drivers/spi/spi-mpc52xx.c mpc52xx_spi_start_transfer(ms); ms 305 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_transfer; ms 314 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_fsm_process(int irq, struct mpc52xx_spi *ms) ms 322 drivers/spi/spi-mpc52xx.c status = in_8(ms->regs + SPI_STATUS); ms 323 drivers/spi/spi-mpc52xx.c data = in_8(ms->regs + SPI_DATA); ms 324 drivers/spi/spi-mpc52xx.c rc = ms->state(irq, ms, status, data); ms 328 drivers/spi/spi-mpc52xx.c schedule_work(&ms->work); ms 336 drivers/spi/spi-mpc52xx.c struct mpc52xx_spi *ms = _ms; ms 337 drivers/spi/spi-mpc52xx.c spin_lock(&ms->lock); ms 338 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsm_process(irq, ms); ms 339 drivers/spi/spi-mpc52xx.c spin_unlock(&ms->lock); ms 348 drivers/spi/spi-mpc52xx.c struct mpc52xx_spi *ms = container_of(work, struct mpc52xx_spi, work); ms 351 drivers/spi/spi-mpc52xx.c spin_lock_irqsave(&ms->lock, flags); ms 352 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsm_process(0, ms); ms 353 drivers/spi/spi-mpc52xx.c spin_unlock_irqrestore(&ms->lock, flags); ms 362 drivers/spi/spi-mpc52xx.c struct mpc52xx_spi *ms = spi_master_get_devdata(spi->master); ms 368 drivers/spi/spi-mpc52xx.c spin_lock_irqsave(&ms->lock, flags); ms 369 drivers/spi/spi-mpc52xx.c list_add_tail(&m->queue, &ms->queue); ms 370 drivers/spi/spi-mpc52xx.c spin_unlock_irqrestore(&ms->lock, flags); ms 371 drivers/spi/spi-mpc52xx.c schedule_work(&ms->work); ms 382 drivers/spi/spi-mpc52xx.c struct mpc52xx_spi *ms; ms 416 drivers/spi/spi-mpc52xx.c master = spi_alloc_master(&op->dev, sizeof *ms); ms 429 drivers/spi/spi-mpc52xx.c ms = spi_master_get_devdata(master); ms 430 drivers/spi/spi-mpc52xx.c ms->master = master; ms 431 drivers/spi/spi-mpc52xx.c ms->regs = regs; ms 432 drivers/spi/spi-mpc52xx.c ms->irq0 = irq_of_parse_and_map(op->dev.of_node, 0); ms 433 drivers/spi/spi-mpc52xx.c ms->irq1 = irq_of_parse_and_map(op->dev.of_node, 1); ms 434 drivers/spi/spi-mpc52xx.c ms->state = mpc52xx_spi_fsmstate_idle; ms 435 drivers/spi/spi-mpc52xx.c ms->ipb_freq = mpc5xxx_get_bus_frequency(op->dev.of_node); ms 436 drivers/spi/spi-mpc52xx.c ms->gpio_cs_count = of_gpio_count(op->dev.of_node); ms 437 drivers/spi/spi-mpc52xx.c if (ms->gpio_cs_count > 0) { ms 438 drivers/spi/spi-mpc52xx.c master->num_chipselect = ms->gpio_cs_count; ms 439 drivers/spi/spi-mpc52xx.c ms->gpio_cs = kmalloc_array(ms->gpio_cs_count, ms 440 drivers/spi/spi-mpc52xx.c sizeof(*ms->gpio_cs), ms 442 drivers/spi/spi-mpc52xx.c if (!ms->gpio_cs) { ms 447 drivers/spi/spi-mpc52xx.c for (i = 0; i < ms->gpio_cs_count; i++) { ms 465 drivers/spi/spi-mpc52xx.c ms->gpio_cs[i] = gpio_cs; ms 469 drivers/spi/spi-mpc52xx.c spin_lock_init(&ms->lock); ms 470 drivers/spi/spi-mpc52xx.c INIT_LIST_HEAD(&ms->queue); ms 471 drivers/spi/spi-mpc52xx.c INIT_WORK(&ms->work, mpc52xx_spi_wq); ms 474 drivers/spi/spi-mpc52xx.c if (ms->irq0 && ms->irq1) { ms 475 drivers/spi/spi-mpc52xx.c rc = request_irq(ms->irq0, mpc52xx_spi_irq, 0, ms 476 drivers/spi/spi-mpc52xx.c "mpc5200-spi-modf", ms); ms 477 drivers/spi/spi-mpc52xx.c rc |= request_irq(ms->irq1, mpc52xx_spi_irq, 0, ms 478 drivers/spi/spi-mpc52xx.c "mpc5200-spi-spif", ms); ms 480 drivers/spi/spi-mpc52xx.c free_irq(ms->irq0, ms); ms 481 drivers/spi/spi-mpc52xx.c free_irq(ms->irq1, ms); ms 482 drivers/spi/spi-mpc52xx.c ms->irq0 = ms->irq1 = 0; ms 486 drivers/spi/spi-mpc52xx.c ms->irq0 = ms->irq1 = 0; ms 489 drivers/spi/spi-mpc52xx.c if (!ms->irq0) ms 497 drivers/spi/spi-mpc52xx.c dev_info(&ms->master->dev, "registered MPC5200 SPI bus\n"); ms 502 drivers/spi/spi-mpc52xx.c dev_err(&ms->master->dev, "initialization failed\n"); ms 505 drivers/spi/spi-mpc52xx.c gpio_free(ms->gpio_cs[i]); ms 507 drivers/spi/spi-mpc52xx.c kfree(ms->gpio_cs); ms 519 drivers/spi/spi-mpc52xx.c struct mpc52xx_spi *ms = spi_master_get_devdata(master); ms 522 drivers/spi/spi-mpc52xx.c free_irq(ms->irq0, ms); ms 523 drivers/spi/spi-mpc52xx.c free_irq(ms->irq1, ms); ms 525 drivers/spi/spi-mpc52xx.c for (i = 0; i < ms->gpio_cs_count; i++) ms 526 drivers/spi/spi-mpc52xx.c gpio_free(ms->gpio_cs[i]); ms 528 drivers/spi/spi-mpc52xx.c kfree(ms->gpio_cs); ms 530 drivers/spi/spi-mpc52xx.c iounmap(ms->regs); ms 448 drivers/spi/spi-s3c64xx.c int ms; ms 451 drivers/spi/spi-s3c64xx.c ms = xfer->len * 8 * 1000 / sdd->cur_speed; ms 452 drivers/spi/spi-s3c64xx.c ms += 10; /* some tolerance */ ms 454 drivers/spi/spi-s3c64xx.c val = msecs_to_jiffies(ms) + 10; ms 494 drivers/spi/spi-s3c64xx.c int ms; ms 497 drivers/spi/spi-s3c64xx.c ms = xfer->len * 8 * 1000 / sdd->cur_speed; ms 498 drivers/spi/spi-s3c64xx.c ms += 10; /* some tolerance */ ms 500 drivers/spi/spi-s3c64xx.c val = msecs_to_loops(ms); ms 527 drivers/spi/spi-s3c64xx.c (loops ? ms : 0)); ms 1063 drivers/spi/spi.c unsigned long long ms = 1; ms 1071 drivers/spi/spi.c ms = 8LL * 1000LL * xfer->len; ms 1072 drivers/spi/spi.c do_div(ms, xfer->speed_hz); ms 1073 drivers/spi/spi.c ms += ms + 200; /* some tolerance */ ms 1075 drivers/spi/spi.c if (ms > UINT_MAX) ms 1076 drivers/spi/spi.c ms = UINT_MAX; ms 1078 drivers/spi/spi.c ms = wait_for_completion_timeout(&ctlr->xfer_completion, ms 1079 drivers/spi/spi.c msecs_to_jiffies(ms)); ms 1081 drivers/spi/spi.c if (ms == 0) { ms 321 drivers/ssb/driver_chipcommon.c u32 ssb_chipco_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms) ms 329 drivers/ssb/driver_chipcommon.c ticks = ssb_chipco_watchdog_timer_set(cc, cc->ticks_per_ms * ms); ms 122 drivers/ssb/driver_extif.c u32 ssb_extif_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms) ms 125 drivers/ssb/driver_extif.c u32 ticks = (SSB_EXTIF_WATCHDOG_CLK / 1000) * ms; ms 198 drivers/ssb/ssb_private.h extern u32 ssb_chipco_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms); ms 221 drivers/ssb/ssb_private.h extern u32 ssb_extif_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms); ms 229 drivers/ssb/ssb_private.h u32 ms) ms 554 drivers/staging/most/configfs.c struct most_sound *ms = container_of(to_configfs_subsystem(group), ms 557 drivers/staging/most/configfs.c list_for_each_entry(most, &ms->soundcard_list, list) { ms 568 drivers/staging/most/configfs.c list_add_tail(&most->list, &ms->soundcard_list); ms 332 drivers/staging/rtl8188eu/include/rtw_mlme.h #define rtw_set_scan_deny(adapter, ms) do {} while (0) ms 565 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h #define set_survey_timer(mlmeext, ms) \ ms 567 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h msecs_to_jiffies(ms)) ms 569 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h #define set_link_timer(mlmeext, ms) \ ms 571 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h msecs_to_jiffies(ms)) ms 211 drivers/staging/rtl8188eu/include/rtw_pwrctrl.h #define _rtw_set_pwr_state_check_timer(pwrctrlpriv, ms) \ ms 213 drivers/staging/rtl8188eu/include/rtw_pwrctrl.h jiffies + msecs_to_jiffies(ms)) ms 1934 drivers/staging/rtl8723bs/core/rtw_mlme.c void rtw_set_scan_deny(struct adapter *adapter, u32 ms) ms 1940 drivers/staging/rtl8723bs/core/rtw_mlme.c _set_timer(&mlmepriv->set_scan_deny_timer, ms); ms 1188 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c inline void rtw_set_ips_deny(struct adapter *padapter, u32 ms) ms 1191 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c pwrpriv->ips_deny_time = jiffies + msecs_to_jiffies(ms); ms 459 drivers/staging/rtl8723bs/include/rtw_mlme.h #define rtw_mlme_set_auto_scan_int(adapter, ms) \ ms 461 drivers/staging/rtl8723bs/include/rtw_mlme.h adapter->mlmepriv.auto_scan_int_ms = ms; \ ms 590 drivers/staging/rtl8723bs/include/rtw_mlme.h void rtw_set_scan_deny(struct adapter *adapter, u32 ms); ms 719 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_survey_timer(mlmeext, ms) \ ms 722 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h _set_timer(&(mlmeext)->survey_timer, (ms)); \ ms 725 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_link_timer(mlmeext, ms) \ ms 728 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h _set_timer(&(mlmeext)->link_timer, (ms)); \ ms 730 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_sa_query_timer(mlmeext, ms) \ ms 732 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h DBG_871X("%s set_sa_query_timer(%p, %d)\n", __func__, (mlmeext), (ms)); \ ms 733 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h _set_timer(&(mlmeext)->sa_query_timer, (ms)); \ ms 315 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h #define _rtw_set_pwr_state_check_timer(pwrctl, ms) \ ms 318 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h _set_timer(&(pwrctl)->pwr_state_check_timer, (ms)); \ ms 352 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h void rtw_set_ips_deny(struct adapter *padapter, u32 ms); ms 302 drivers/thermal/intel/x86_pkg_temp_thermal.c unsigned long ms = msecs_to_jiffies(notify_delay_ms); ms 304 drivers/thermal/intel/x86_pkg_temp_thermal.c schedule_delayed_work_on(cpu, work, ms); ms 248 drivers/tty/hvc/hvsi_lib.c static void maybe_msleep(unsigned long ms) ms 252 drivers/tty/hvc/hvsi_lib.c mdelay(ms); ms 254 drivers/tty/hvc/hvsi_lib.c msleep(ms); ms 1525 drivers/tty/mxser.c struct mxser_mstatus ms, __user *msu = argp; ms 1530 drivers/tty/mxser.c memset(&ms, 0, sizeof(ms)); ms 1539 drivers/tty/mxser.c ms.cflag = ip->normal_termios.c_cflag; ms 1541 drivers/tty/mxser.c ms.cflag = tty->termios.c_cflag; ms 1547 drivers/tty/mxser.c ms.dcd = 1; ms 1549 drivers/tty/mxser.c ms.dsr = 1; ms 1551 drivers/tty/mxser.c ms.cts = 1; ms 1554 drivers/tty/mxser.c if (copy_to_user(msu, &ms, sizeof(ms))) ms 1401 drivers/tty/serial/sunsu.c const char *ms = of_get_property(ap, "mouse", NULL); ms 1419 drivers/tty/serial/sunsu.c if (ms) { ms 1420 drivers/tty/serial/sunsu.c match = of_find_node_by_path(ms); ms 147 drivers/usb/class/cdc-acm.c #define acm_send_break(acm, ms) \ ms 148 drivers/usb/class/cdc-acm.c acm_ctrl_msg(acm, USB_CDC_REQ_SEND_BREAK, ms, NULL, 0) ms 398 drivers/usb/dwc2/core.c s64 ms; ms 407 drivers/usb/dwc2/core.c ms = ktime_to_ms(ktime_sub(end, start)); ms 409 drivers/usb/dwc2/core.c if (ms >= (s64)timeout) { ms 185 drivers/usb/early/ehci-dbgp.c static inline void dbgp_mdelay(int ms) ms 189 drivers/usb/early/ehci-dbgp.c while (ms--) { ms 581 drivers/usb/storage/scsiglue.c unsigned short ms; ms 583 drivers/usb/storage/scsiglue.c if (sscanf(buf, "%hu", &ms) > 0) { ms 584 drivers/usb/storage/scsiglue.c blk_queue_max_hw_sectors(sdev->request_queue, ms); ms 42 drivers/video/fbdev/amba-clcd.c static inline void clcdfb_sleep(unsigned int ms) ms 45 drivers/video/fbdev/amba-clcd.c mdelay(ms); ms 47 drivers/video/fbdev/amba-clcd.c msleep(ms); ms 284 drivers/video/fbdev/aty/radeon_base.c void _radeon_msleep(struct radeonfb_info *rinfo, unsigned long ms) ms 287 drivers/video/fbdev/aty/radeon_base.c mdelay(ms); ms 289 drivers/video/fbdev/aty/radeon_base.c msleep(ms); ms 374 drivers/video/fbdev/aty/radeonfb.h void _radeon_msleep(struct radeonfb_info *rinfo, unsigned long ms); ms 477 drivers/video/fbdev/aty/radeonfb.h #define radeon_msleep(ms) _radeon_msleep(rinfo,ms) ms 511 drivers/w1/masters/ds2490.c int ms = del<<4; ms 520 drivers/w1/masters/ds2490.c if (delay == 0 || ms == dev->spu_sleep) ms 527 drivers/w1/masters/ds2490.c dev->spu_sleep = ms; ms 87 fs/dlm/lock.c struct dlm_message *ms); ms 88 fs/dlm/lock.c static int receive_extralen(struct dlm_message *ms); ms 1509 fs/dlm/lock.c struct dlm_message *ms) ms 1548 fs/dlm/lock.c is_overlap_cancel(lkb) && ms && !ms->m_result) { ms 1566 fs/dlm/lock.c lkb->lkb_id, ms ? ms->m_header.h_nodeid : 0, lkb->lkb_remid, ms 1607 fs/dlm/lock.c static int remove_from_waiters_ms(struct dlm_lkb *lkb, struct dlm_message *ms) ms 1612 fs/dlm/lock.c if (ms->m_flags != DLM_IFL_STUB_MS) ms 1614 fs/dlm/lock.c error = _remove_from_waiters(lkb, ms->m_type, ms); ms 1615 fs/dlm/lock.c if (ms->m_flags != DLM_IFL_STUB_MS) ms 2026 fs/dlm/lock.c struct dlm_message *ms) ms 2038 fs/dlm/lock.c int len = receive_extralen(ms); ms 2041 fs/dlm/lock.c memcpy(lkb->lkb_lvbptr, ms->m_extra, len); ms 2042 fs/dlm/lock.c lkb->lkb_lvbseq = ms->m_lvbseq; ms 2133 fs/dlm/lock.c struct dlm_message *ms) ms 2135 fs/dlm/lock.c set_lvb_lock_pc(r, lkb, ms); ms 2171 fs/dlm/lock.c static void munge_altmode(struct dlm_lkb *lkb, struct dlm_message *ms) ms 2173 fs/dlm/lock.c if (ms->m_type != DLM_MSG_REQUEST_REPLY && ms 2174 fs/dlm/lock.c ms->m_type != DLM_MSG_GRANT) { ms 2176 fs/dlm/lock.c lkb->lkb_id, ms->m_type); ms 3532 fs/dlm/lock.c struct dlm_message *ms; ms 3546 fs/dlm/lock.c ms = (struct dlm_message *) mb; ms 3548 fs/dlm/lock.c ms->m_header.h_version = (DLM_HEADER_MAJOR | DLM_HEADER_MINOR); ms 3549 fs/dlm/lock.c ms->m_header.h_lockspace = ls->ls_global_id; ms 3550 fs/dlm/lock.c ms->m_header.h_nodeid = dlm_our_nodeid(); ms 3551 fs/dlm/lock.c ms->m_header.h_length = mb_len; ms 3552 fs/dlm/lock.c ms->m_header.h_cmd = DLM_MSG; ms 3554 fs/dlm/lock.c ms->m_type = mstype; ms 3557 fs/dlm/lock.c *ms_ret = ms; ms 3591 fs/dlm/lock.c static int send_message(struct dlm_mhandle *mh, struct dlm_message *ms) ms 3593 fs/dlm/lock.c dlm_message_out(ms); ms 3599 fs/dlm/lock.c struct dlm_message *ms) ms 3601 fs/dlm/lock.c ms->m_nodeid = lkb->lkb_nodeid; ms 3602 fs/dlm/lock.c ms->m_pid = lkb->lkb_ownpid; ms 3603 fs/dlm/lock.c ms->m_lkid = lkb->lkb_id; ms 3604 fs/dlm/lock.c ms->m_remid = lkb->lkb_remid; ms 3605 fs/dlm/lock.c ms->m_exflags = lkb->lkb_exflags; ms 3606 fs/dlm/lock.c ms->m_sbflags = lkb->lkb_sbflags; ms 3607 fs/dlm/lock.c ms->m_flags = lkb->lkb_flags; ms 3608 fs/dlm/lock.c ms->m_lvbseq = lkb->lkb_lvbseq; ms 3609 fs/dlm/lock.c ms->m_status = lkb->lkb_status; ms 3610 fs/dlm/lock.c ms->m_grmode = lkb->lkb_grmode; ms 3611 fs/dlm/lock.c ms->m_rqmode = lkb->lkb_rqmode; ms 3612 fs/dlm/lock.c ms->m_hash = r->res_hash; ms 3618 fs/dlm/lock.c ms->m_asts |= DLM_CB_BAST; ms 3620 fs/dlm/lock.c ms->m_asts |= DLM_CB_CAST; ms 3625 fs/dlm/lock.c switch (ms->m_type) { ms 3628 fs/dlm/lock.c memcpy(ms->m_extra, r->res_name, r->res_length); ms 3637 fs/dlm/lock.c memcpy(ms->m_extra, lkb->lkb_lvbptr, r->res_ls->ls_lvblen); ms 3644 fs/dlm/lock.c struct dlm_message *ms; ms 3654 fs/dlm/lock.c error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh); ms 3658 fs/dlm/lock.c send_args(r, lkb, ms); ms 3660 fs/dlm/lock.c error = send_message(mh, ms); ms 3709 fs/dlm/lock.c struct dlm_message *ms; ms 3715 fs/dlm/lock.c error = create_message(r, lkb, to_nodeid, DLM_MSG_GRANT, &ms, &mh); ms 3719 fs/dlm/lock.c send_args(r, lkb, ms); ms 3721 fs/dlm/lock.c ms->m_result = 0; ms 3723 fs/dlm/lock.c error = send_message(mh, ms); ms 3730 fs/dlm/lock.c struct dlm_message *ms; ms 3736 fs/dlm/lock.c error = create_message(r, NULL, to_nodeid, DLM_MSG_BAST, &ms, &mh); ms 3740 fs/dlm/lock.c send_args(r, lkb, ms); ms 3742 fs/dlm/lock.c ms->m_bastmode = mode; ms 3744 fs/dlm/lock.c error = send_message(mh, ms); ms 3751 fs/dlm/lock.c struct dlm_message *ms; ms 3761 fs/dlm/lock.c error = create_message(r, NULL, to_nodeid, DLM_MSG_LOOKUP, &ms, &mh); ms 3765 fs/dlm/lock.c send_args(r, lkb, ms); ms 3767 fs/dlm/lock.c error = send_message(mh, ms); ms 3779 fs/dlm/lock.c struct dlm_message *ms; ms 3785 fs/dlm/lock.c error = create_message(r, NULL, to_nodeid, DLM_MSG_REMOVE, &ms, &mh); ms 3789 fs/dlm/lock.c memcpy(ms->m_extra, r->res_name, r->res_length); ms 3790 fs/dlm/lock.c ms->m_hash = r->res_hash; ms 3792 fs/dlm/lock.c error = send_message(mh, ms); ms 3800 fs/dlm/lock.c struct dlm_message *ms; ms 3806 fs/dlm/lock.c error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh); ms 3810 fs/dlm/lock.c send_args(r, lkb, ms); ms 3812 fs/dlm/lock.c ms->m_result = rv; ms 3814 fs/dlm/lock.c error = send_message(mh, ms); ms 3843 fs/dlm/lock.c struct dlm_message *ms; ms 3847 fs/dlm/lock.c error = create_message(r, NULL, nodeid, DLM_MSG_LOOKUP_REPLY, &ms, &mh); ms 3851 fs/dlm/lock.c ms->m_lkid = ms_in->m_lkid; ms 3852 fs/dlm/lock.c ms->m_result = rv; ms 3853 fs/dlm/lock.c ms->m_nodeid = ret_nodeid; ms 3855 fs/dlm/lock.c error = send_message(mh, ms); ms 3864 fs/dlm/lock.c static void receive_flags(struct dlm_lkb *lkb, struct dlm_message *ms) ms 3866 fs/dlm/lock.c lkb->lkb_exflags = ms->m_exflags; ms 3867 fs/dlm/lock.c lkb->lkb_sbflags = ms->m_sbflags; ms 3869 fs/dlm/lock.c (ms->m_flags & 0x0000FFFF); ms 3872 fs/dlm/lock.c static void receive_flags_reply(struct dlm_lkb *lkb, struct dlm_message *ms) ms 3874 fs/dlm/lock.c if (ms->m_flags == DLM_IFL_STUB_MS) ms 3877 fs/dlm/lock.c lkb->lkb_sbflags = ms->m_sbflags; ms 3879 fs/dlm/lock.c (ms->m_flags & 0x0000FFFF); ms 3882 fs/dlm/lock.c static int receive_extralen(struct dlm_message *ms) ms 3884 fs/dlm/lock.c return (ms->m_header.h_length - sizeof(struct dlm_message)); ms 3888 fs/dlm/lock.c struct dlm_message *ms) ms 3897 fs/dlm/lock.c len = receive_extralen(ms); ms 3900 fs/dlm/lock.c memcpy(lkb->lkb_lvbptr, ms->m_extra, len); ms 3916 fs/dlm/lock.c struct dlm_message *ms) ms 3918 fs/dlm/lock.c lkb->lkb_nodeid = ms->m_header.h_nodeid; ms 3919 fs/dlm/lock.c lkb->lkb_ownpid = ms->m_pid; ms 3920 fs/dlm/lock.c lkb->lkb_remid = ms->m_lkid; ms 3922 fs/dlm/lock.c lkb->lkb_rqmode = ms->m_rqmode; ms 3924 fs/dlm/lock.c lkb->lkb_bastfn = (ms->m_asts & DLM_CB_BAST) ? &fake_bastfn : NULL; ms 3925 fs/dlm/lock.c lkb->lkb_astfn = (ms->m_asts & DLM_CB_CAST) ? &fake_astfn : NULL; ms 3938 fs/dlm/lock.c struct dlm_message *ms) ms 3943 fs/dlm/lock.c if (receive_lvb(ls, lkb, ms)) ms 3946 fs/dlm/lock.c lkb->lkb_rqmode = ms->m_rqmode; ms 3947 fs/dlm/lock.c lkb->lkb_lvbseq = ms->m_lvbseq; ms 3953 fs/dlm/lock.c struct dlm_message *ms) ms 3955 fs/dlm/lock.c if (receive_lvb(ls, lkb, ms)) ms 3963 fs/dlm/lock.c static void setup_stub_lkb(struct dlm_ls *ls, struct dlm_message *ms) ms 3966 fs/dlm/lock.c lkb->lkb_nodeid = ms->m_header.h_nodeid; ms 3967 fs/dlm/lock.c lkb->lkb_remid = ms->m_lkid; ms 3973 fs/dlm/lock.c static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms) ms 3975 fs/dlm/lock.c int from = ms->m_header.h_nodeid; ms 3978 fs/dlm/lock.c switch (ms->m_type) { ms 4009 fs/dlm/lock.c ms->m_type, from, lkb->lkb_id, lkb->lkb_remid, ms 4017 fs/dlm/lock.c struct dlm_message *ms; ms 4057 fs/dlm/lock.c dir_nodeid, DLM_MSG_REMOVE, &ms, &mh); ms 4061 fs/dlm/lock.c memcpy(ms->m_extra, name, len); ms 4062 fs/dlm/lock.c ms->m_hash = hash; ms 4064 fs/dlm/lock.c send_message(mh, ms); ms 4072 fs/dlm/lock.c static int receive_request(struct dlm_ls *ls, struct dlm_message *ms) ms 4079 fs/dlm/lock.c from_nodeid = ms->m_header.h_nodeid; ms 4085 fs/dlm/lock.c receive_flags(lkb, ms); ms 4087 fs/dlm/lock.c error = receive_request_args(ls, lkb, ms); ms 4099 fs/dlm/lock.c namelen = receive_extralen(ms); ms 4101 fs/dlm/lock.c error = find_rsb(ls, ms->m_extra, namelen, from_nodeid, ms 4152 fs/dlm/lock.c ms->m_lkid, from_nodeid, error); ms 4156 fs/dlm/lock.c send_repeat_remove(ls, ms->m_extra, namelen); ms 4160 fs/dlm/lock.c setup_stub_lkb(ls, ms); ms 4165 fs/dlm/lock.c static int receive_convert(struct dlm_ls *ls, struct dlm_message *ms) ms 4171 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4175 fs/dlm/lock.c if (lkb->lkb_remid != ms->m_lkid) { ms 4179 fs/dlm/lock.c ms->m_header.h_nodeid, ms->m_lkid); ms 4190 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4194 fs/dlm/lock.c receive_flags(lkb, ms); ms 4196 fs/dlm/lock.c error = receive_convert_args(ls, lkb, ms); ms 4215 fs/dlm/lock.c setup_stub_lkb(ls, ms); ms 4220 fs/dlm/lock.c static int receive_unlock(struct dlm_ls *ls, struct dlm_message *ms) ms 4226 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4230 fs/dlm/lock.c if (lkb->lkb_remid != ms->m_lkid) { ms 4233 fs/dlm/lock.c ms->m_header.h_nodeid, ms->m_lkid); ms 4244 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4248 fs/dlm/lock.c receive_flags(lkb, ms); ms 4250 fs/dlm/lock.c error = receive_unlock_args(ls, lkb, ms); ms 4266 fs/dlm/lock.c setup_stub_lkb(ls, ms); ms 4271 fs/dlm/lock.c static int receive_cancel(struct dlm_ls *ls, struct dlm_message *ms) ms 4277 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4281 fs/dlm/lock.c receive_flags(lkb, ms); ms 4288 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4302 fs/dlm/lock.c setup_stub_lkb(ls, ms); ms 4307 fs/dlm/lock.c static int receive_grant(struct dlm_ls *ls, struct dlm_message *ms) ms 4313 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4322 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4326 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4328 fs/dlm/lock.c munge_altmode(lkb, ms); ms 4329 fs/dlm/lock.c grant_lock_pc(r, lkb, ms); ms 4338 fs/dlm/lock.c static int receive_bast(struct dlm_ls *ls, struct dlm_message *ms) ms 4344 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4353 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4357 fs/dlm/lock.c queue_bast(r, lkb, ms->m_bastmode); ms 4358 fs/dlm/lock.c lkb->lkb_highbast = ms->m_bastmode; ms 4366 fs/dlm/lock.c static void receive_lookup(struct dlm_ls *ls, struct dlm_message *ms) ms 4370 fs/dlm/lock.c from_nodeid = ms->m_header.h_nodeid; ms 4373 fs/dlm/lock.c len = receive_extralen(ms); ms 4375 fs/dlm/lock.c error = dlm_master_lookup(ls, from_nodeid, ms->m_extra, len, 0, ms 4380 fs/dlm/lock.c receive_request(ls, ms); ms 4383 fs/dlm/lock.c send_lookup_reply(ls, ms, ret_nodeid, error); ms 4386 fs/dlm/lock.c static void receive_remove(struct dlm_ls *ls, struct dlm_message *ms) ms 4393 fs/dlm/lock.c from_nodeid = ms->m_header.h_nodeid; ms 4395 fs/dlm/lock.c len = receive_extralen(ms); ms 4403 fs/dlm/lock.c dir_nodeid = dlm_hash2nodeid(ls, ms->m_hash); ms 4420 fs/dlm/lock.c memcpy(name, ms->m_extra, len); ms 4474 fs/dlm/lock.c static void receive_purge(struct dlm_ls *ls, struct dlm_message *ms) ms 4476 fs/dlm/lock.c do_purge(ls, ms->m_nodeid, ms->m_pid); ms 4479 fs/dlm/lock.c static int receive_request_reply(struct dlm_ls *ls, struct dlm_message *ms) ms 4484 fs/dlm/lock.c int from_nodeid = ms->m_header.h_nodeid; ms 4486 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4494 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4502 fs/dlm/lock.c lkb->lkb_id, from_nodeid, ms->m_lkid, ms->m_result); ms 4516 fs/dlm/lock.c result = ms->m_result; ms 4529 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4530 fs/dlm/lock.c lkb->lkb_remid = ms->m_lkid; ms 4532 fs/dlm/lock.c munge_altmode(lkb, ms); ms 4537 fs/dlm/lock.c grant_lock_pc(r, lkb, ms); ms 4600 fs/dlm/lock.c struct dlm_message *ms) ms 4603 fs/dlm/lock.c switch (ms->m_result) { ms 4610 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4617 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4627 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4630 fs/dlm/lock.c grant_lock_pc(r, lkb, ms); ms 4636 fs/dlm/lock.c lkb->lkb_id, ms->m_header.h_nodeid, ms->m_lkid, ms 4637 fs/dlm/lock.c ms->m_result); ms 4643 fs/dlm/lock.c static void _receive_convert_reply(struct dlm_lkb *lkb, struct dlm_message *ms) ms 4651 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4656 fs/dlm/lock.c error = remove_from_waiters_ms(lkb, ms); ms 4660 fs/dlm/lock.c __receive_convert_reply(r, lkb, ms); ms 4666 fs/dlm/lock.c static int receive_convert_reply(struct dlm_ls *ls, struct dlm_message *ms) ms 4671 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4675 fs/dlm/lock.c _receive_convert_reply(lkb, ms); ms 4680 fs/dlm/lock.c static void _receive_unlock_reply(struct dlm_lkb *lkb, struct dlm_message *ms) ms 4688 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4693 fs/dlm/lock.c error = remove_from_waiters_ms(lkb, ms); ms 4699 fs/dlm/lock.c switch (ms->m_result) { ms 4701 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4709 fs/dlm/lock.c lkb->lkb_id, ms->m_result); ms 4716 fs/dlm/lock.c static int receive_unlock_reply(struct dlm_ls *ls, struct dlm_message *ms) ms 4721 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4725 fs/dlm/lock.c _receive_unlock_reply(lkb, ms); ms 4730 fs/dlm/lock.c static void _receive_cancel_reply(struct dlm_lkb *lkb, struct dlm_message *ms) ms 4738 fs/dlm/lock.c error = validate_message(lkb, ms); ms 4743 fs/dlm/lock.c error = remove_from_waiters_ms(lkb, ms); ms 4749 fs/dlm/lock.c switch (ms->m_result) { ms 4751 fs/dlm/lock.c receive_flags_reply(lkb, ms); ms 4759 fs/dlm/lock.c lkb->lkb_id, ms->m_result); ms 4766 fs/dlm/lock.c static int receive_cancel_reply(struct dlm_ls *ls, struct dlm_message *ms) ms 4771 fs/dlm/lock.c error = find_lkb(ls, ms->m_remid, &lkb); ms 4775 fs/dlm/lock.c _receive_cancel_reply(lkb, ms); ms 4780 fs/dlm/lock.c static void receive_lookup_reply(struct dlm_ls *ls, struct dlm_message *ms) ms 4787 fs/dlm/lock.c error = find_lkb(ls, ms->m_lkid, &lkb); ms 4789 fs/dlm/lock.c log_error(ls, "receive_lookup_reply no lkid %x", ms->m_lkid); ms 4804 fs/dlm/lock.c ret_nodeid = ms->m_nodeid; ms 4816 fs/dlm/lock.c lkb->lkb_id, ms->m_header.h_nodeid, ret_nodeid, ms 4829 fs/dlm/lock.c lkb->lkb_id, ms->m_header.h_nodeid); ms 4858 fs/dlm/lock.c static void _receive_message(struct dlm_ls *ls, struct dlm_message *ms, ms 4863 fs/dlm/lock.c if (!dlm_is_member(ls, ms->m_header.h_nodeid)) { ms 4865 fs/dlm/lock.c ms->m_type, ms->m_header.h_nodeid, ms->m_lkid, ms 4866 fs/dlm/lock.c ms->m_remid, ms->m_result); ms 4870 fs/dlm/lock.c switch (ms->m_type) { ms 4875 fs/dlm/lock.c error = receive_request(ls, ms); ms 4879 fs/dlm/lock.c error = receive_convert(ls, ms); ms 4883 fs/dlm/lock.c error = receive_unlock(ls, ms); ms 4888 fs/dlm/lock.c error = receive_cancel(ls, ms); ms 4894 fs/dlm/lock.c error = receive_request_reply(ls, ms); ms 4898 fs/dlm/lock.c error = receive_convert_reply(ls, ms); ms 4902 fs/dlm/lock.c error = receive_unlock_reply(ls, ms); ms 4906 fs/dlm/lock.c error = receive_cancel_reply(ls, ms); ms 4913 fs/dlm/lock.c error = receive_grant(ls, ms); ms 4918 fs/dlm/lock.c error = receive_bast(ls, ms); ms 4924 fs/dlm/lock.c receive_lookup(ls, ms); ms 4928 fs/dlm/lock.c receive_remove(ls, ms); ms 4934 fs/dlm/lock.c receive_lookup_reply(ls, ms); ms 4940 fs/dlm/lock.c receive_purge(ls, ms); ms 4944 fs/dlm/lock.c log_error(ls, "unknown message type %d", ms->m_type); ms 4960 fs/dlm/lock.c ms->m_type, ms->m_remid, ms->m_header.h_nodeid, ms 4961 fs/dlm/lock.c ms->m_lkid, saved_seq); ms 4964 fs/dlm/lock.c ms->m_type, ms->m_remid, ms->m_header.h_nodeid, ms 4965 fs/dlm/lock.c ms->m_lkid, saved_seq); ms 4967 fs/dlm/lock.c if (ms->m_type == DLM_MSG_CONVERT) ms 4968 fs/dlm/lock.c dlm_dump_rsb_hash(ls, ms->m_hash); ms 4974 fs/dlm/lock.c ms->m_type, ms->m_header.h_nodeid, ms 4975 fs/dlm/lock.c ms->m_lkid, ms->m_remid, saved_seq); ms 4987 fs/dlm/lock.c static void dlm_receive_message(struct dlm_ls *ls, struct dlm_message *ms, ms 4996 fs/dlm/lock.c ms->m_type, nodeid); ms 5000 fs/dlm/lock.c dlm_add_requestqueue(ls, nodeid, ms); ms 5003 fs/dlm/lock.c _receive_message(ls, ms, 0); ms 5010 fs/dlm/lock.c void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms, ms 5013 fs/dlm/lock.c _receive_message(ls, ms, saved_seq); ms 6274 fs/dlm/lock.c struct dlm_message *ms; ms 6279 fs/dlm/lock.c DLM_MSG_PURGE, &ms, &mh); ms 6282 fs/dlm/lock.c ms->m_nodeid = nodeid; ms 6283 fs/dlm/lock.c ms->m_pid = pid; ms 6285 fs/dlm/lock.c return send_message(mh, ms); ms 17 fs/dlm/lock.h void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms, ms 32 fs/dlm/requestqueue.c void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms) ms 35 fs/dlm/requestqueue.c int length = ms->m_header.h_length - sizeof(struct dlm_message); ms 45 fs/dlm/requestqueue.c memcpy(&e->request, ms, ms->m_header.h_length); ms 66 fs/dlm/requestqueue.c struct dlm_message *ms; ms 80 fs/dlm/requestqueue.c ms = &e->request; ms 84 fs/dlm/requestqueue.c ms->m_type, ms->m_header.h_nodeid, ms 85 fs/dlm/requestqueue.c ms->m_lkid, ms->m_remid, ms->m_result, ms 128 fs/dlm/requestqueue.c static int purge_request(struct dlm_ls *ls, struct dlm_message *ms, int nodeid) ms 130 fs/dlm/requestqueue.c uint32_t type = ms->m_type; ms 155 fs/dlm/requestqueue.c struct dlm_message *ms; ms 160 fs/dlm/requestqueue.c ms = &e->request; ms 162 fs/dlm/requestqueue.c if (purge_request(ls, ms, e->nodeid)) { ms 14 fs/dlm/requestqueue.h void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms); ms 84 fs/dlm/util.c void dlm_message_out(struct dlm_message *ms) ms 86 fs/dlm/util.c header_out(&ms->m_header); ms 88 fs/dlm/util.c ms->m_type = cpu_to_le32(ms->m_type); ms 89 fs/dlm/util.c ms->m_nodeid = cpu_to_le32(ms->m_nodeid); ms 90 fs/dlm/util.c ms->m_pid = cpu_to_le32(ms->m_pid); ms 91 fs/dlm/util.c ms->m_lkid = cpu_to_le32(ms->m_lkid); ms 92 fs/dlm/util.c ms->m_remid = cpu_to_le32(ms->m_remid); ms 93 fs/dlm/util.c ms->m_parent_lkid = cpu_to_le32(ms->m_parent_lkid); ms 94 fs/dlm/util.c ms->m_parent_remid = cpu_to_le32(ms->m_parent_remid); ms 95 fs/dlm/util.c ms->m_exflags = cpu_to_le32(ms->m_exflags); ms 96 fs/dlm/util.c ms->m_sbflags = cpu_to_le32(ms->m_sbflags); ms 97 fs/dlm/util.c ms->m_flags = cpu_to_le32(ms->m_flags); ms 98 fs/dlm/util.c ms->m_lvbseq = cpu_to_le32(ms->m_lvbseq); ms 99 fs/dlm/util.c ms->m_hash = cpu_to_le32(ms->m_hash); ms 100 fs/dlm/util.c ms->m_status = cpu_to_le32(ms->m_status); ms 101 fs/dlm/util.c ms->m_grmode = cpu_to_le32(ms->m_grmode); ms 102 fs/dlm/util.c ms->m_rqmode = cpu_to_le32(ms->m_rqmode); ms 103 fs/dlm/util.c ms->m_bastmode = cpu_to_le32(ms->m_bastmode); ms 104 fs/dlm/util.c ms->m_asts = cpu_to_le32(ms->m_asts); ms 105 fs/dlm/util.c ms->m_result = cpu_to_le32(to_dlm_errno(ms->m_result)); ms 108 fs/dlm/util.c void dlm_message_in(struct dlm_message *ms) ms 110 fs/dlm/util.c header_in(&ms->m_header); ms 112 fs/dlm/util.c ms->m_type = le32_to_cpu(ms->m_type); ms 113 fs/dlm/util.c ms->m_nodeid = le32_to_cpu(ms->m_nodeid); ms 114 fs/dlm/util.c ms->m_pid = le32_to_cpu(ms->m_pid); ms 115 fs/dlm/util.c ms->m_lkid = le32_to_cpu(ms->m_lkid); ms 116 fs/dlm/util.c ms->m_remid = le32_to_cpu(ms->m_remid); ms 117 fs/dlm/util.c ms->m_parent_lkid = le32_to_cpu(ms->m_parent_lkid); ms 118 fs/dlm/util.c ms->m_parent_remid = le32_to_cpu(ms->m_parent_remid); ms 119 fs/dlm/util.c ms->m_exflags = le32_to_cpu(ms->m_exflags); ms 120 fs/dlm/util.c ms->m_sbflags = le32_to_cpu(ms->m_sbflags); ms 121 fs/dlm/util.c ms->m_flags = le32_to_cpu(ms->m_flags); ms 122 fs/dlm/util.c ms->m_lvbseq = le32_to_cpu(ms->m_lvbseq); ms 123 fs/dlm/util.c ms->m_hash = le32_to_cpu(ms->m_hash); ms 124 fs/dlm/util.c ms->m_status = le32_to_cpu(ms->m_status); ms 125 fs/dlm/util.c ms->m_grmode = le32_to_cpu(ms->m_grmode); ms 126 fs/dlm/util.c ms->m_rqmode = le32_to_cpu(ms->m_rqmode); ms 127 fs/dlm/util.c ms->m_bastmode = le32_to_cpu(ms->m_bastmode); ms 128 fs/dlm/util.c ms->m_asts = le32_to_cpu(ms->m_asts); ms 129 fs/dlm/util.c ms->m_result = from_dlm_errno(le32_to_cpu(ms->m_result)); ms 14 fs/dlm/util.h void dlm_message_out(struct dlm_message *ms); ms 15 fs/dlm/util.h void dlm_message_in(struct dlm_message *ms); ms 1797 fs/eventpoll.c static inline struct timespec64 ep_set_mstimeout(long ms) ms 1800 fs/eventpoll.c .tv_sec = ms / MSEC_PER_SEC, ms 1801 fs/eventpoll.c .tv_nsec = NSEC_PER_MSEC * (ms % MSEC_PER_SEC), ms 176 fs/isofs/inode.c isofs_hashi_common(const struct dentry *dentry, struct qstr *qstr, int ms) ms 185 fs/isofs/inode.c if (ms) { ms 205 fs/isofs/inode.c const struct qstr *name, int ms, int ci) ms 212 fs/isofs/inode.c if (ms) { ms 248 fs/isofs/inode.c isofs_hash_common(const struct dentry *dentry, struct qstr *qstr, int ms) ms 255 fs/isofs/inode.c if (ms) { ms 119 fs/minix/inode.c struct minix_super_block * ms; ms 122 fs/minix/inode.c ms = sbi->s_ms; ms 126 fs/minix/inode.c if (ms->s_state & MINIX_VALID_FS || ms 131 fs/minix/inode.c ms->s_state = sbi->s_mount_state; ms 136 fs/minix/inode.c sbi->s_mount_state = ms->s_state; ms 137 fs/minix/inode.c ms->s_state &= ~MINIX_VALID_FS; ms 157 fs/minix/inode.c struct minix_super_block *ms; ms 178 fs/minix/inode.c ms = (struct minix_super_block *) bh->b_data; ms 179 fs/minix/inode.c sbi->s_ms = ms; ms 181 fs/minix/inode.c sbi->s_mount_state = ms->s_state; ms 182 fs/minix/inode.c sbi->s_ninodes = ms->s_ninodes; ms 183 fs/minix/inode.c sbi->s_nzones = ms->s_nzones; ms 184 fs/minix/inode.c sbi->s_imap_blocks = ms->s_imap_blocks; ms 185 fs/minix/inode.c sbi->s_zmap_blocks = ms->s_zmap_blocks; ms 186 fs/minix/inode.c sbi->s_firstdatazone = ms->s_firstdatazone; ms 187 fs/minix/inode.c sbi->s_log_zone_size = ms->s_log_zone_size; ms 188 fs/minix/inode.c sbi->s_max_size = ms->s_max_size; ms 189 fs/minix/inode.c s->s_magic = ms->s_magic; ms 202 fs/minix/inode.c sbi->s_nzones = ms->s_zones; ms 208 fs/minix/inode.c sbi->s_nzones = ms->s_zones; ms 295 fs/minix/inode.c ms->s_state &= ~MINIX_VALID_FS; ms 37 include/asm-generic/mmiowb.h struct mmiowb_state *ms = __mmiowb_state(); ms 38 include/asm-generic/mmiowb.h ms->mmiowb_pending = ms->nesting_count; ms 43 include/asm-generic/mmiowb.h struct mmiowb_state *ms = __mmiowb_state(); ms 44 include/asm-generic/mmiowb.h ms->nesting_count++; ms 49 include/asm-generic/mmiowb.h struct mmiowb_state *ms = __mmiowb_state(); ms 51 include/asm-generic/mmiowb.h if (unlikely(ms->mmiowb_pending)) { ms 52 include/asm-generic/mmiowb.h ms->mmiowb_pending = 0; ms 56 include/asm-generic/mmiowb.h ms->nesting_count--; ms 148 include/linux/ceph/libceph.h #define from_msgr(ms) container_of(ms, struct ceph_client, msgr) ms 270 include/linux/ktime.h static inline ktime_t ms_to_ktime(u64 ms) ms 272 include/linux/ktime.h return ms * NSEC_PER_MSEC; ms 354 include/linux/memory_hotplug.h extern void sparse_remove_section(struct mem_section *ms, ms 1230 include/linux/mmzone.h static inline unsigned long *section_to_usemap(struct mem_section *ms) ms 1232 include/linux/mmzone.h return ms->usage->pageblock_flags; ms 1245 include/linux/mmzone.h extern unsigned long __section_nr(struct mem_section *ms); ms 1331 include/linux/mmzone.h static inline int pfn_section_valid(struct mem_section *ms, unsigned long pfn) ms 1335 include/linux/mmzone.h return test_bit(idx, ms->usage->subsection_map); ms 1338 include/linux/mmzone.h static inline int pfn_section_valid(struct mem_section *ms, unsigned long pfn) ms 1347 include/linux/mmzone.h struct mem_section *ms; ms 1351 include/linux/mmzone.h ms = __nr_to_section(pfn_to_section_nr(pfn)); ms 1352 include/linux/mmzone.h if (!valid_section(ms)) ms 1358 include/linux/mmzone.h return early_section(ms) || pfn_section_valid(ms, pfn); ms 43 include/linux/module_signature.h int mod_check_sig(const struct module_signature *ms, size_t file_len, ms 301 include/linux/nvme.h __le16 ms; ms 226 include/linux/regulator/consumer.h int regulator_disable_deferred(struct regulator *regulator, int ms); ms 416 include/linux/regulator/consumer.h int ms) ms 948 include/net/ip_vs.h struct ipvs_master_sync_state *ms; ms 501 include/sound/wavefront.h wavefront_multisample ms; ms 154 include/uapi/drm/via_drm.h __u32 ms; ms 220 include/video/imx-ipu-v3.h int ipu_idmac_wait_busy(struct ipuv3_channel *channel, int ms); ms 21 kernel/module_signature.c int mod_check_sig(const struct module_signature *ms, size_t file_len, ms 24 kernel/module_signature.c if (be32_to_cpu(ms->sig_len) >= file_len - sizeof(*ms)) ms 27 kernel/module_signature.c if (ms->id_type != PKEY_ID_PKCS7) { ms 33 kernel/module_signature.c if (ms->algo != 0 || ms 34 kernel/module_signature.c ms->hash != 0 || ms 35 kernel/module_signature.c ms->signer_len != 0 || ms 36 kernel/module_signature.c ms->key_id_len != 0 || ms 37 kernel/module_signature.c ms->__pad[0] != 0 || ms 38 kernel/module_signature.c ms->__pad[1] != 0 || ms 39 kernel/module_signature.c ms->__pad[2] != 0) { ms 22 kernel/module_signing.c struct module_signature ms; ms 28 kernel/module_signing.c if (modlen <= sizeof(ms)) ms 31 kernel/module_signing.c memcpy(&ms, mod + (modlen - sizeof(ms)), sizeof(ms)); ms 33 kernel/module_signing.c ret = mod_check_sig(&ms, modlen, info->name); ms 37 kernel/module_signing.c sig_len = be32_to_cpu(ms.sig_len); ms 38 kernel/module_signing.c modlen -= sig_len + sizeof(ms); ms 169 mm/memory_hotplug.c struct mem_section *ms; ms 174 mm/memory_hotplug.c ms = __nr_to_section(section_nr); ms 177 mm/memory_hotplug.c memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr); ms 191 mm/memory_hotplug.c usage = ms->usage; ms 204 mm/memory_hotplug.c struct mem_section *ms; ms 209 mm/memory_hotplug.c ms = __nr_to_section(section_nr); ms 211 mm/memory_hotplug.c memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr); ms 215 mm/memory_hotplug.c usage = ms->usage; ms 499 mm/memory_hotplug.c struct mem_section *ms = __nr_to_section(pfn_to_section_nr(pfn)); ms 501 mm/memory_hotplug.c if (WARN_ON_ONCE(!valid_section(ms))) ms 504 mm/memory_hotplug.c sparse_remove_section(ms, pfn, nr_pages, map_offset, altmap); ms 275 mm/page_ext.c struct mem_section *ms; ms 278 mm/page_ext.c ms = __pfn_to_section(pfn); ms 279 mm/page_ext.c if (!ms || !ms->page_ext) ms 281 mm/page_ext.c base = get_entry(ms->page_ext, pfn); ms 283 mm/page_ext.c ms->page_ext = NULL; ms 114 mm/sparse.c unsigned long __section_nr(struct mem_section *ms) ms 124 mm/sparse.c if ((ms >= root) && (ms < (root + SECTIONS_PER_ROOT))) ms 130 mm/sparse.c return (root_nr * SECTIONS_PER_ROOT) + (ms - root); ms 133 mm/sparse.c unsigned long __section_nr(struct mem_section *ms) ms 135 mm/sparse.c return (unsigned long)(ms - mem_section[0]); ms 191 mm/sparse.c static void section_mark_present(struct mem_section *ms) ms 193 mm/sparse.c unsigned long section_nr = __section_nr(ms); ms 198 mm/sparse.c ms->section_mem_map |= SECTION_MARKED_PRESENT; ms 240 mm/sparse.c struct mem_section *ms; ms 245 mm/sparse.c ms = __nr_to_section(nr); ms 246 mm/sparse.c subsection_mask_set(ms->usage->subsection_map, pfn, pfns); ms 279 mm/sparse.c struct mem_section *ms; ms 284 mm/sparse.c ms = __nr_to_section(section); ms 285 mm/sparse.c if (!ms->section_mem_map) { ms 286 mm/sparse.c ms->section_mem_map = sparse_encode_early_nid(nid) | ms 288 mm/sparse.c section_mark_present(ms); ms 333 mm/sparse.c static void __meminit sparse_init_one_section(struct mem_section *ms, ms 337 mm/sparse.c ms->section_mem_map &= ~SECTION_MAP_MASK; ms 338 mm/sparse.c ms->section_mem_map |= sparse_encode_mem_map(mem_map, pnum) ms 340 mm/sparse.c ms->usage = usage; ms 566 mm/sparse.c struct mem_section *ms; ms 570 mm/sparse.c ms = __nr_to_section(pnum); ms 571 mm/sparse.c ms->section_mem_map = 0; ms 615 mm/sparse.c struct mem_section *ms; ms 621 mm/sparse.c ms = __nr_to_section(section_nr); ms 622 mm/sparse.c ms->section_mem_map |= SECTION_IS_ONLINE; ms 634 mm/sparse.c struct mem_section *ms; ms 643 mm/sparse.c ms = __nr_to_section(section_nr); ms 644 mm/sparse.c ms->section_mem_map &= ~SECTION_IS_ONLINE; ms 742 mm/sparse.c struct mem_section *ms = __pfn_to_section(pfn); ms 743 mm/sparse.c bool section_is_early = early_section(ms); ms 746 mm/sparse.c unsigned long *subsection_map = ms->usage ms 747 mm/sparse.c ? &ms->usage->subsection_map[0] : NULL; ms 787 mm/sparse.c if (!PageReserved(virt_to_page(ms->usage))) { ms 788 mm/sparse.c kfree(ms->usage); ms 789 mm/sparse.c ms->usage = NULL; ms 791 mm/sparse.c memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr); ms 797 mm/sparse.c ms->section_mem_map &= ~SECTION_HAS_MEM_MAP; ms 806 mm/sparse.c ms->section_mem_map = (unsigned long)NULL; ms 813 mm/sparse.c struct mem_section *ms = __pfn_to_section(pfn); ms 821 mm/sparse.c if (!ms->usage) { ms 825 mm/sparse.c ms->usage = usage; ms 827 mm/sparse.c subsection_map = &ms->usage->subsection_map[0]; ms 839 mm/sparse.c ms->usage = NULL; ms 851 mm/sparse.c if (nr_pages < PAGES_PER_SECTION && early_section(ms)) ms 881 mm/sparse.c struct mem_section *ms; ms 899 mm/sparse.c ms = __nr_to_section(section_nr); ms 901 mm/sparse.c section_mark_present(ms); ms 906 mm/sparse.c sparse_init_one_section(ms, section_nr, memmap, ms->usage, 0); ms 938 mm/sparse.c void sparse_remove_section(struct mem_section *ms, unsigned long pfn, ms 3574 net/mac80211/mlme.c u32 tu, ms; ms 3576 net/mac80211/mlme.c ms = tu * 1024 / 1000; ms 3579 net/mac80211/mlme.c mgmt->sa, tu, ms); ms 3580 net/mac80211/mlme.c assoc_data->timeout = jiffies + msecs_to_jiffies(ms); ms 3582 net/mac80211/mlme.c if (ms > IEEE80211_ASSOC_TIMEOUT) ms 61 net/mac80211/rc80211_minstrel_debugfs.c struct minstrel_debugfs_info *ms; ms 65 net/mac80211/rc80211_minstrel_debugfs.c ms = kmalloc(2048, GFP_KERNEL); ms 66 net/mac80211/rc80211_minstrel_debugfs.c if (!ms) ms 69 net/mac80211/rc80211_minstrel_debugfs.c file->private_data = ms; ms 70 net/mac80211/rc80211_minstrel_debugfs.c p = ms->buf; ms 112 net/mac80211/rc80211_minstrel_debugfs.c ms->len = p - ms->buf; ms 114 net/mac80211/rc80211_minstrel_debugfs.c WARN_ON(ms->len + sizeof(*ms) > 2048); ms 123 net/mac80211/rc80211_minstrel_debugfs.c struct minstrel_debugfs_info *ms; ms 127 net/mac80211/rc80211_minstrel_debugfs.c ms = kmalloc(2048, GFP_KERNEL); ms 128 net/mac80211/rc80211_minstrel_debugfs.c if (!ms) ms 131 net/mac80211/rc80211_minstrel_debugfs.c file->private_data = ms; ms 132 net/mac80211/rc80211_minstrel_debugfs.c p = ms->buf; ms 167 net/mac80211/rc80211_minstrel_debugfs.c ms->len = p - ms->buf; ms 169 net/mac80211/rc80211_minstrel_debugfs.c WARN_ON(ms->len + sizeof(*ms) > 2048); ms 18 net/mac80211/rc80211_minstrel_ht_debugfs.c struct minstrel_debugfs_info *ms; ms 20 net/mac80211/rc80211_minstrel_ht_debugfs.c ms = file->private_data; ms 21 net/mac80211/rc80211_minstrel_ht_debugfs.c return simple_read_from_buffer(buf, len, ppos, ms->buf, ms->len); ms 125 net/mac80211/rc80211_minstrel_ht_debugfs.c struct minstrel_debugfs_info *ms; ms 137 net/mac80211/rc80211_minstrel_ht_debugfs.c ms = kmalloc(32768, GFP_KERNEL); ms 138 net/mac80211/rc80211_minstrel_ht_debugfs.c if (!ms) ms 141 net/mac80211/rc80211_minstrel_ht_debugfs.c file->private_data = ms; ms 142 net/mac80211/rc80211_minstrel_ht_debugfs.c p = ms->buf; ms 164 net/mac80211/rc80211_minstrel_ht_debugfs.c ms->len = p - ms->buf; ms 165 net/mac80211/rc80211_minstrel_ht_debugfs.c WARN_ON(ms->len + sizeof(*ms) > 32768); ms 275 net/mac80211/rc80211_minstrel_ht_debugfs.c struct minstrel_debugfs_info *ms; ms 287 net/mac80211/rc80211_minstrel_ht_debugfs.c ms = kmalloc(32768, GFP_KERNEL); ms 289 net/mac80211/rc80211_minstrel_ht_debugfs.c if (!ms) ms 292 net/mac80211/rc80211_minstrel_ht_debugfs.c file->private_data = ms; ms 294 net/mac80211/rc80211_minstrel_ht_debugfs.c p = ms->buf; ms 302 net/mac80211/rc80211_minstrel_ht_debugfs.c ms->len = p - ms->buf; ms 303 net/mac80211/rc80211_minstrel_ht_debugfs.c WARN_ON(ms->len + sizeof(*ms) > 32768); ms 307 net/netfilter/ipvs/ip_vs_sync.c sb_dequeue(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms) ms 312 net/netfilter/ipvs/ip_vs_sync.c if (list_empty(&ms->sync_queue)) { ms 316 net/netfilter/ipvs/ip_vs_sync.c sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff, ms 319 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_len--; ms 320 net/netfilter/ipvs/ip_vs_sync.c if (!ms->sync_queue_len) ms 321 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_delay = 0; ms 366 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms) ms 368 net/netfilter/ipvs/ip_vs_sync.c struct ip_vs_sync_buff *sb = ms->sync_buff; ms 372 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_len < sysctl_sync_qlen_max(ipvs)) { ms 373 net/netfilter/ipvs/ip_vs_sync.c if (!ms->sync_queue_len) ms 374 net/netfilter/ipvs/ip_vs_sync.c schedule_delayed_work(&ms->master_wakeup_work, ms 376 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_len++; ms 377 net/netfilter/ipvs/ip_vs_sync.c list_add_tail(&sb->list, &ms->sync_queue); ms 378 net/netfilter/ipvs/ip_vs_sync.c if ((++ms->sync_queue_delay) == IPVS_SYNC_WAKEUP_RATE) { ms 379 net/netfilter/ipvs/ip_vs_sync.c int id = (int)(ms - ipvs->ms); ms 393 net/netfilter/ipvs/ip_vs_sync.c get_curr_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms, ms 399 net/netfilter/ipvs/ip_vs_sync.c sb = ms->sync_buff; ms 401 net/netfilter/ipvs/ip_vs_sync.c ms->sync_buff = NULL; ms 549 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms; ms 569 net/netfilter/ipvs/ip_vs_sync.c ms = &ipvs->ms[id]; ms 570 net/netfilter/ipvs/ip_vs_sync.c buff = ms->sync_buff; ms 577 net/netfilter/ipvs/ip_vs_sync.c sb_queue_tail(ipvs, ms); ms 578 net/netfilter/ipvs/ip_vs_sync.c ms->sync_buff = NULL; ms 589 net/netfilter/ipvs/ip_vs_sync.c ms->sync_buff = buff; ms 638 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms; ms 672 net/netfilter/ipvs/ip_vs_sync.c ms = &ipvs->ms[id]; ms 691 net/netfilter/ipvs/ip_vs_sync.c buff = ms->sync_buff; ms 697 net/netfilter/ipvs/ip_vs_sync.c sb_queue_tail(ipvs, ms); ms 698 net/netfilter/ipvs/ip_vs_sync.c ms->sync_buff = NULL; ms 711 net/netfilter/ipvs/ip_vs_sync.c ms->sync_buff = buff; ms 1635 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms = ms 1638 net/netfilter/ipvs/ip_vs_sync.c struct netns_ipvs *ipvs = ms->ipvs; ms 1641 net/netfilter/ipvs/ip_vs_sync.c if (ms->sync_queue_len && ms 1642 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_delay < IPVS_SYNC_WAKEUP_RATE) { ms 1643 net/netfilter/ipvs/ip_vs_sync.c int id = (int)(ms - ipvs->ms); ms 1645 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_delay = IPVS_SYNC_WAKEUP_RATE; ms 1653 net/netfilter/ipvs/ip_vs_sync.c next_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms) ms 1657 net/netfilter/ipvs/ip_vs_sync.c sb = sb_dequeue(ipvs, ms); ms 1661 net/netfilter/ipvs/ip_vs_sync.c return get_curr_sync_buff(ipvs, ms, IPVS_SYNC_FLUSH_TIME); ms 1668 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms = &ipvs->ms[tinfo->id]; ms 1677 net/netfilter/ipvs/ip_vs_sync.c sb = next_sync_buff(ipvs, ms); ms 1703 net/netfilter/ipvs/ip_vs_sync.c while ((sb = sb_dequeue(ipvs, ms))) ms 1708 net/netfilter/ipvs/ip_vs_sync.c sb = get_curr_sync_buff(ipvs, ms, 0); ms 1818 net/netfilter/ipvs/ip_vs_sync.c if (ipvs->ms) ms 1838 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms; ms 1841 net/netfilter/ipvs/ip_vs_sync.c ipvs->ms = kcalloc(count, sizeof(ipvs->ms[0]), GFP_KERNEL); ms 1842 net/netfilter/ipvs/ip_vs_sync.c if (!ipvs->ms) ms 1844 net/netfilter/ipvs/ip_vs_sync.c ms = ipvs->ms; ms 1845 net/netfilter/ipvs/ip_vs_sync.c for (id = 0; id < count; id++, ms++) { ms 1846 net/netfilter/ipvs/ip_vs_sync.c INIT_LIST_HEAD(&ms->sync_queue); ms 1847 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_len = 0; ms 1848 net/netfilter/ipvs/ip_vs_sync.c ms->sync_queue_delay = 0; ms 1849 net/netfilter/ipvs/ip_vs_sync.c INIT_DELAYED_WORK(&ms->master_wakeup_work, ms 1851 net/netfilter/ipvs/ip_vs_sync.c ms->ipvs = ipvs; ms 1914 net/netfilter/ipvs/ip_vs_sync.c kfree(ipvs->ms); ms 1915 net/netfilter/ipvs/ip_vs_sync.c ipvs->ms = NULL; ms 1954 net/netfilter/ipvs/ip_vs_sync.c if (!ipvs->ms) ms 1972 net/netfilter/ipvs/ip_vs_sync.c struct ipvs_master_sync_state *ms = &ipvs->ms[id]; ms 1978 net/netfilter/ipvs/ip_vs_sync.c cancel_delayed_work_sync(&ms->master_wakeup_work); ms 1983 net/netfilter/ipvs/ip_vs_sync.c kfree(ipvs->ms); ms 1984 net/netfilter/ipvs/ip_vs_sync.c ipvs->ms = NULL; ms 3280 net/netfilter/nf_tables_api.c u64 ms = be64_to_cpu(nla_get_be64(nla)); ms 3284 net/netfilter/nf_tables_api.c if (ms >= max) ms 3287 net/netfilter/nf_tables_api.c ms *= NSEC_PER_MSEC; ms 3288 net/netfilter/nf_tables_api.c *result = nsecs_to_jiffies64(ms); ms 262 sound/core/seq/seq_midi.c struct seq_midisynth *msynth, *ms; ms 325 sound/core/seq/seq_midi.c ms = &msynth[p]; ms 327 sound/core/seq/seq_midi.c if (snd_seq_midisynth_new(ms, card, device, p) < 0) ms 371 sound/core/seq/seq_midi.c pcallbacks.private_data = ms; ms 382 sound/core/seq/seq_midi.c ms->seq_client = client->seq_client; ms 383 sound/core/seq/seq_midi.c ms->seq_port = port->addr.port; ms 1216 sound/isa/wavefront/wavefront_synth.c num_samples = (1<<(header->hdr.ms.NumberOfSamples&7)); ms 1217 sound/isa/wavefront/wavefront_synth.c msample_hdr[2] = (unsigned char) header->hdr.ms.NumberOfSamples; ms 1221 sound/isa/wavefront/wavefront_synth.c header->hdr.ms.NumberOfSamples, ms 1226 sound/isa/wavefront/wavefront_synth.c i, header->hdr.ms.SampleNumber[i]); ms 1227 sound/isa/wavefront/wavefront_synth.c munge_int32 (header->hdr.ms.SampleNumber[i], ms 1269 sound/isa/wavefront/wavefront_synth.c header->hdr.ms.NumberOfSamples = log_ns[0]; ms 1293 sound/isa/wavefront/wavefront_synth.c header->hdr.ms.SampleNumber[i] = ms 1297 sound/isa/wavefront/wavefront_synth.c i, header->hdr.ms.SampleNumber[i]); ms 2625 sound/soc/codecs/wm8994.c int ms = 0; ms 2650 sound/soc/codecs/wm8994.c ms = WM8994_AIF1_MSTR; ms 2736 sound/soc/codecs/wm8994.c ms); ms 14 tools/perf/arch/arm64/annotate/instructions.c struct map_symbol *ms __maybe_unused) ms 5 tools/perf/arch/s390/annotate/instructions.c struct map_symbol *ms) ms 8 tools/perf/arch/s390/annotate/instructions.c struct map *map = ms->map; ms 58 tools/perf/arch/s390/annotate/instructions.c struct map_symbol *ms __maybe_unused) ms 304 tools/perf/builtin-annotate.c return symbol__tty_annotate(he->ms.sym, he->ms.map, evsel, &ann->opts); ms 306 tools/perf/builtin-annotate.c return symbol__tty_annotate2(he->ms.sym, he->ms.map, evsel, &ann->opts); ms 320 tools/perf/builtin-annotate.c if (he->ms.sym == NULL || he->ms.map->dso->annotate_warned) ms 324 tools/perf/builtin-annotate.c (strcmp(he->ms.sym->name, ann->sym_hist_filter) != 0)) ms 327 tools/perf/builtin-annotate.c notes = symbol__annotation(he->ms.sym); ms 621 tools/perf/builtin-diff.c if (!he->ms.map || !he->ms.sym) ms 624 tools/perf/builtin-diff.c notes = symbol__annotation(he->ms.sym); ms 632 tools/perf/builtin-diff.c for (unsigned int i = 0; i < symbol__size(he->ms.sym); i++) { ms 641 tools/perf/builtin-diff.c init_block_info(bi, he->ms.sym, &ch[i], i); ms 1352 tools/perf/builtin-diff.c start_line = map__srcline(he->ms.map, bi->sym->start + bi->start, ms 1353 tools/perf/builtin-diff.c he->ms.sym); ms 1355 tools/perf/builtin-diff.c end_line = map__srcline(he->ms.map, bi->sym->start + bi->end, ms 1356 tools/perf/builtin-diff.c he->ms.sym); ms 643 tools/perf/builtin-report.c struct symbol *sym = he->ms.sym; ms 648 tools/perf/builtin-report.c symbol__annotate2(sym, he->ms.map, evsel, ms 2762 tools/perf/builtin-sched.c if (chain->ms.sym && chain->ms.sym->ignore) ms 115 tools/perf/builtin-top.c if (!he || !he->ms.sym) ms 120 tools/perf/builtin-top.c sym = he->ms.sym; ms 121 tools/perf/builtin-top.c map = he->ms.map; ms 160 tools/perf/builtin-top.c struct symbol *sym = he->ms.sym; ms 197 tools/perf/builtin-top.c struct symbol *sym = he->ms.sym; ms 202 tools/perf/builtin-top.c top->sym_filter_entry->ms.sym != sym))) ms 221 tools/perf/builtin-top.c if (err == -ERANGE && !he->ms.map->erange_warned) ms 222 tools/perf/builtin-top.c ui__warn_map_erange(he->ms.map, sym, ip); ms 246 tools/perf/builtin-top.c symbol = he->ms.sym; ms 259 tools/perf/builtin-top.c more = symbol__annotate_printf(symbol, he->ms.map, top->sym_evsel, &top->annotation_opts); ms 412 tools/perf/builtin-top.c if (n->ms.sym && !strcmp(buf, n->ms.sym->name)) { ms 434 tools/perf/builtin-top.c struct symbol *sym = top->sym_filter_entry->ms.sym; ms 184 tools/perf/tests/hists_common.c he->ms.map->dso->short_name, ms 185 tools/perf/tests/hists_common.c he->ms.sym->name, he->stat.period); ms 211 tools/perf/tests/hists_common.c he->ms.map->dso->short_name, ms 212 tools/perf/tests/hists_common.c he->ms.sym->name, he->stat.period, ms 153 tools/perf/tests/hists_cumulate.c #define DSO(he) (he->ms.map->dso->short_name) ms 154 tools/perf/tests/hists_cumulate.c #define SYM(he) (he->ms.sym->name) ms 158 tools/perf/tests/hists_cumulate.c #define CDSO(cl) (cl->ms.map->dso->short_name) ms 159 tools/perf/tests/hists_cumulate.c #define CSYM(cl) (cl->ms.sym->name) ms 164 tools/perf/tests/hists_link.c he->thread, he->ms.map, he->ms.sym)) { ms 216 tools/perf/tests/hists_link.c he->thread, he->ms.map, he->ms.sym) && ms 219 tools/perf/tests/hists_link.c he->thread, he->ms.map, he->ms.sym)) { ms 119 tools/perf/tests/hists_output.c #define DSO(he) (he->ms.map->dso->short_name) ms 120 tools/perf/tests/hists_output.c #define SYM(he) (he->ms.sym->name) ms 42 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->priv; ms 43 tools/perf/ui/browsers/annotate.c return symbol__annotation(ms->sym); ms 153 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = ab->b.priv; ms 154 tools/perf/ui/browsers/annotate.c struct symbol *sym = ms->sym; ms 305 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; ms 306 tools/perf/ui/browsers/annotate.c struct symbol *sym = ms->sym; ms 413 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; ms 434 tools/perf/ui/browsers/annotate.c symbol__tui_annotate(dl->ops.target.sym, ms->map, evsel, hbt, browser->opts); ms 435 tools/perf/ui/browsers/annotate.c sym_title(ms->sym, ms->map, title, sizeof(title), browser->opts->percent_type); ms 611 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->priv; ms 612 tools/perf/ui/browsers/annotate.c struct symbol *sym = ms->sym; ms 618 tools/perf/ui/browsers/annotate.c sym_title(sym, ms->map, symbol_dso, sizeof(symbol_dso), ab->opts->percent_type); ms 665 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; ms 666 tools/perf/ui/browsers/annotate.c struct symbol *sym = ms->sym; ms 667 tools/perf/ui/browsers/annotate.c struct annotation *notes = symbol__annotation(ms->sym); ms 831 tools/perf/ui/browsers/annotate.c map_symbol__annotation_dump(ms, evsel, browser->opts); ms 873 tools/perf/ui/browsers/annotate.c int map_symbol__tui_annotate(struct map_symbol *ms, struct evsel *evsel, ms 877 tools/perf/ui/browsers/annotate.c return symbol__tui_annotate(ms->sym, ms->map, evsel, hbt, opts); ms 888 tools/perf/ui/browsers/annotate.c return map_symbol__tui_annotate(&he->ms, evsel, hbt, opts); ms 897 tools/perf/ui/browsers/annotate.c struct map_symbol ms = { ms 908 tools/perf/ui/browsers/annotate.c .priv = &ms, ms 396 tools/perf/ui/browsers/hists.c struct map_symbol *ms = browser->selection; ms 397 tools/perf/ui/browsers/hists.c struct callchain_list *cl = container_of(ms, struct callchain_list, ms); ms 400 tools/perf/ui/browsers/hists.c if (!he || !ms) ms 403 tools/perf/ui/browsers/hists.c if (ms == &he->ms) ms 745 tools/perf/ui/browsers/hists.c bool show_annotated = browser->show_dso && chain->ms.sym && symbol__annotation(chain->ms.sym)->src; ms 750 tools/perf/ui/browsers/hists.c browser->selection = &chain->ms; ms 1259 tools/perf/ui/browsers/hists.c browser->selection = &entry->ms; ms 1372 tools/perf/ui/browsers/hists.c browser->selection = &entry->ms; ms 2359 tools/perf/ui/browsers/hists.c struct map_symbol ms; ms 2379 tools/perf/ui/browsers/hists.c notes = symbol__annotation(act->ms.sym); ms 2384 tools/perf/ui/browsers/hists.c err = map_symbol__tui_annotate(&act->ms, evsel, browser->hbt, ms 2411 tools/perf/ui/browsers/hists.c act->ms.map = map; ms 2412 tools/perf/ui/browsers/hists.c act->ms.sym = sym; ms 2482 tools/perf/ui/browsers/hists.c struct map *map = act->ms.map; ms 2517 tools/perf/ui/browsers/hists.c act->ms.map = map; ms 2526 tools/perf/ui/browsers/hists.c map__browse(act->ms.map); ms 2540 tools/perf/ui/browsers/hists.c act->ms.map = map; ms 2556 tools/perf/ui/browsers/hists.c else if (act->ms.sym) ms 2557 tools/perf/ui/browsers/hists.c len += strlen(act->ms.sym->name); ms 2566 tools/perf/ui/browsers/hists.c } else if (act->ms.sym) { ms 2568 tools/perf/ui/browsers/hists.c act->ms.sym->name); ms 2622 tools/perf/ui/browsers/hists.c act->ms.sym = sym; ms 2945 tools/perf/ui/browsers/hists.c actions->ms.map = browser->selection->map; ms 2946 tools/perf/ui/browsers/hists.c actions->ms.sym = browser->selection->sym; ms 2953 tools/perf/ui/browsers/hists.c actions->ms.map = map; ms 2984 tools/perf/ui/browsers/hists.c actions->ms.sym = NULL; ms 3057 tools/perf/ui/browsers/hists.c actions->ms.map = map; ms 245 tools/perf/ui/gtk/annotate.c return symbol__gtk_annotate(he->ms.sym, he->ms.map, evsel, hbt); ms 105 tools/perf/ui/stdio/hist.c rem_hits.ms.sym = rem_sq_bracket; ms 860 tools/perf/ui/stdio/hist.c if (h->ms.map == NULL && verbose > 1) { ms 240 tools/perf/util/annotate.c static int call__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms) ms 243 tools/perf/util/annotate.c struct map *map = ms->map; ms 329 tools/perf/util/annotate.c static int jump__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms) ms 331 tools/perf/util/annotate.c struct map *map = ms->map; ms 332 tools/perf/util/annotate.c struct symbol *sym = ms->sym; ms 477 tools/perf/util/annotate.c static int lock__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms) ms 492 tools/perf/util/annotate.c ops->locked.ins.ops->parse(arch, ops->locked.ops, ms) < 0) ms 535 tools/perf/util/annotate.c static int mov__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms __maybe_unused) ms 596 tools/perf/util/annotate.c static int dec__parse(struct arch *arch __maybe_unused, struct ins_operands *ops, struct map_symbol *ms __maybe_unused) ms 1097 tools/perf/util/annotate.c return symbol__inc_addr_samples(he->ms.sym, he->ms.map, evsel, ip, sample); ms 1100 tools/perf/util/annotate.c static void disasm_line__init_ins(struct disasm_line *dl, struct arch *arch, struct map_symbol *ms) ms 1107 tools/perf/util/annotate.c if (dl->ins.ops->parse && dl->ins.ops->parse(arch, &dl->ops, ms) < 0) ms 1142 tools/perf/util/annotate.c struct map_symbol ms; ms 1225 tools/perf/util/annotate.c disasm_line__init_ins(dl, args->arch, &args->ms); ms 1492 tools/perf/util/annotate.c struct map *map = args->ms.map; ms 1539 tools/perf/util/annotate.c args->ms.sym = sym; ms 1721 tools/perf/util/annotate.c struct map *map = args->ms.map; ms 1825 tools/perf/util/annotate.c args->ms.sym = sym; ms 1836 tools/perf/util/annotate.c args->ms.sym = sym; ms 1863 tools/perf/util/annotate.c struct map *map = args->ms.map; ms 2108 tools/perf/util/annotate.c args.ms.map = map; ms 2109 tools/perf/util/annotate.c args.ms.sym = sym; ms 2431 tools/perf/util/annotate.c int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel, ms 2440 tools/perf/util/annotate.c if (asprintf(&filename, "%s.annotation", ms->sym->name) < 0) ms 2453 tools/perf/util/annotate.c ms->sym->name, ms->map->dso->long_name, ev_name); ms 2454 tools/perf/util/annotate.c symbol__annotate_fprintf2(ms->sym, fp, opts); ms 60 tools/perf/util/annotate.h int (*parse)(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms); ms 391 tools/perf/util/annotate.h int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel, ms 585 tools/perf/util/callchain.c call->ms.sym = cursor_node->sym; ms 586 tools/perf/util/callchain.c call->ms.map = map__get(cursor_node->map); ms 645 tools/perf/util/callchain.c map__zput(call->ms.map); ms 723 tools/perf/util/callchain.c if (node->sym && cnode->ms.sym) { ms 730 tools/perf/util/callchain.c if (cnode->ms.sym->inlined || node->sym->inlined) { ms 731 tools/perf/util/callchain.c match = match_chain_strings(cnode->ms.sym->name, ms 736 tools/perf/util/callchain.c match = match_chain_dso_addresses(cnode->ms.map, cnode->ms.sym->start, ms 745 tools/perf/util/callchain.c match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->map, node->ip); ms 1008 tools/perf/util/callchain.c list->ms.map, list->ms.sym, ms 1011 tools/perf/util/callchain.c map__zput(list->ms.map); ms 1154 tools/perf/util/callchain.c if (cl->ms.sym) { ms 1155 tools/perf/util/callchain.c const char *inlined = cl->ms.sym->inlined ? " (inlined)" : ""; ms 1159 tools/perf/util/callchain.c cl->ms.sym->name, cl->srcline, ms 1163 tools/perf/util/callchain.c cl->ms.sym->name, inlined); ms 1169 tools/perf/util/callchain.c cl->ms.map ? ms 1170 tools/perf/util/callchain.c cl->ms.map->dso->short_name : ms 1462 tools/perf/util/callchain.c map__zput(list->ms.map); ms 1468 tools/perf/util/callchain.c map__zput(list->ms.map); ms 1532 tools/perf/util/callchain.c map__get(new->ms.map); ms 1553 tools/perf/util/callchain.c map__zput(chain->ms.map); ms 119 tools/perf/util/callchain.h struct map_symbol ms; ms 88 tools/perf/util/hist.c if (h->ms.sym) { ms 89 tools/perf/util/hist.c symlen = h->ms.sym->namelen + 4; ms 103 tools/perf/util/hist.c if (h->ms.map) { ms 104 tools/perf/util/hist.c len = dso__name_len(h->ms.map->dso); ms 428 tools/perf/util/hist.c map__get(he->ms.map); ms 501 tools/perf/util/hist.c map__zput(he->ms.map); ms 613 tools/perf/util/hist.c if (he->ms.map != entry->ms.map) { ms 614 tools/perf/util/hist.c map__put(he->ms.map); ms 615 tools/perf/util/hist.c he->ms.map = map__get(entry->ms.map); ms 691 tools/perf/util/hist.c .ms = { ms 1064 tools/perf/util/hist.c .ms = { ms 1244 tools/perf/util/hist.c map__zput(he->ms.map); ms 2044 tools/perf/util/hist.c (he->ms.map == NULL || he->ms.map->dso != hists->dso_filter)) { ms 2068 tools/perf/util/hist.c (!he->ms.sym || strstr(he->ms.sym->name, ms 456 tools/perf/util/hist.h int map_symbol__tui_annotate(struct map_symbol *ms, struct evsel *evsel, ms 489 tools/perf/util/hist.h static inline int map_symbol__tui_annotate(struct map_symbol *ms __maybe_unused, ms 191 tools/perf/util/sort.c return _sort__dso_cmp(right->ms.map, left->ms.map); ms 209 tools/perf/util/sort.c return _hist_entry__dso_snprintf(he->ms.map, bf, size, width); ms 219 tools/perf/util/sort.c return dso && (!he->ms.map || he->ms.map->dso != dso); ms 265 tools/perf/util/sort.c if (!left->ms.sym && !right->ms.sym) ms 278 tools/perf/util/sort.c return _sort__sym_cmp(left->ms.sym, right->ms.sym); ms 284 tools/perf/util/sort.c if (!left->ms.sym || !right->ms.sym) ms 285 tools/perf/util/sort.c return cmp_null(left->ms.sym, right->ms.sym); ms 287 tools/perf/util/sort.c return strcmp(right->ms.sym->name, left->ms.sym->name); ms 328 tools/perf/util/sort.c return _hist_entry__sym_snprintf(he->ms.map, he->ms.sym, he->ip, ms 339 tools/perf/util/sort.c return sym && (!he->ms.sym || !strstr(he->ms.sym->name, sym)); ms 355 tools/perf/util/sort.c return map__srcline(he->ms.map, he->ip, he->ms.sym); ms 448 tools/perf/util/sort.c struct symbol *sym = he->ms.sym; ms 502 tools/perf/util/sort.c struct map *map = e->ms.map; ms 508 tools/perf/util/sort.c e->ms.sym, false, true, true, e->ip); ms 1587 tools/perf/util/sort.c return _sort__sym_size_cmp(right->ms.sym, left->ms.sym); ms 1602 tools/perf/util/sort.c return _hist_entry__sym_size_snprintf(he->ms.sym, bf, size, width); ms 1626 tools/perf/util/sort.c return _sort__dso_size_cmp(right->ms.map, left->ms.map); ms 1642 tools/perf/util/sort.c return _hist_entry__dso_size_snprintf(he->ms.map, bf, size, width); ms 96 tools/perf/util/sort.h struct map_symbol ms;