ms                197 arch/arm/mach-pxa/include/mach/regs-lcd.h #define SMART_DELAY(ms)		(SMART_CMD_DELAY | ((ms) & 0xff))
ms                896 arch/ia64/kernel/mca.c 	const pal_min_state_area_t *ms = sos->pal_min_state;
ms                903 arch/ia64/kernel/mca.c 		regs->cr_iip = ms->pmsa_iip;
ms                904 arch/ia64/kernel/mca.c 		regs->cr_ipsr = ms->pmsa_ipsr;
ms                905 arch/ia64/kernel/mca.c 		regs->cr_ifs = ms->pmsa_ifs;
ms                907 arch/ia64/kernel/mca.c 		regs->cr_iip = ms->pmsa_xip;
ms                908 arch/ia64/kernel/mca.c 		regs->cr_ipsr = ms->pmsa_xpsr;
ms                909 arch/ia64/kernel/mca.c 		regs->cr_ifs = ms->pmsa_xfs;
ms                911 arch/ia64/kernel/mca.c 		sos->iip = ms->pmsa_iip;
ms                912 arch/ia64/kernel/mca.c 		sos->ipsr = ms->pmsa_ipsr;
ms                913 arch/ia64/kernel/mca.c 		sos->ifs = ms->pmsa_ifs;
ms                915 arch/ia64/kernel/mca.c 	regs->pr = ms->pmsa_pr;
ms                916 arch/ia64/kernel/mca.c 	regs->b0 = ms->pmsa_br0;
ms                917 arch/ia64/kernel/mca.c 	regs->ar_rsc = ms->pmsa_rsc;
ms                918 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[1-1], ms->pmsa_nat_bits, &regs->r1, nat);
ms                919 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[2-1], ms->pmsa_nat_bits, &regs->r2, nat);
ms                920 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[3-1], ms->pmsa_nat_bits, &regs->r3, nat);
ms                921 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[8-1], ms->pmsa_nat_bits, &regs->r8, nat);
ms                922 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[9-1], ms->pmsa_nat_bits, &regs->r9, nat);
ms                923 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[10-1], ms->pmsa_nat_bits, &regs->r10, nat);
ms                924 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[11-1], ms->pmsa_nat_bits, &regs->r11, nat);
ms                925 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[12-1], ms->pmsa_nat_bits, &regs->r12, nat);
ms                926 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[13-1], ms->pmsa_nat_bits, &regs->r13, nat);
ms                927 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[14-1], ms->pmsa_nat_bits, &regs->r14, nat);
ms                928 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[15-1], ms->pmsa_nat_bits, &regs->r15, nat);
ms                930 arch/ia64/kernel/mca.c 		bank = ms->pmsa_bank1_gr;
ms                932 arch/ia64/kernel/mca.c 		bank = ms->pmsa_bank0_gr;
ms                933 arch/ia64/kernel/mca.c 	copy_reg(&bank[16-16], ms->pmsa_nat_bits, &regs->r16, nat);
ms                934 arch/ia64/kernel/mca.c 	copy_reg(&bank[17-16], ms->pmsa_nat_bits, &regs->r17, nat);
ms                935 arch/ia64/kernel/mca.c 	copy_reg(&bank[18-16], ms->pmsa_nat_bits, &regs->r18, nat);
ms                936 arch/ia64/kernel/mca.c 	copy_reg(&bank[19-16], ms->pmsa_nat_bits, &regs->r19, nat);
ms                937 arch/ia64/kernel/mca.c 	copy_reg(&bank[20-16], ms->pmsa_nat_bits, &regs->r20, nat);
ms                938 arch/ia64/kernel/mca.c 	copy_reg(&bank[21-16], ms->pmsa_nat_bits, &regs->r21, nat);
ms                939 arch/ia64/kernel/mca.c 	copy_reg(&bank[22-16], ms->pmsa_nat_bits, &regs->r22, nat);
ms                940 arch/ia64/kernel/mca.c 	copy_reg(&bank[23-16], ms->pmsa_nat_bits, &regs->r23, nat);
ms                941 arch/ia64/kernel/mca.c 	copy_reg(&bank[24-16], ms->pmsa_nat_bits, &regs->r24, nat);
ms                942 arch/ia64/kernel/mca.c 	copy_reg(&bank[25-16], ms->pmsa_nat_bits, &regs->r25, nat);
ms                943 arch/ia64/kernel/mca.c 	copy_reg(&bank[26-16], ms->pmsa_nat_bits, &regs->r26, nat);
ms                944 arch/ia64/kernel/mca.c 	copy_reg(&bank[27-16], ms->pmsa_nat_bits, &regs->r27, nat);
ms                945 arch/ia64/kernel/mca.c 	copy_reg(&bank[28-16], ms->pmsa_nat_bits, &regs->r28, nat);
ms                946 arch/ia64/kernel/mca.c 	copy_reg(&bank[29-16], ms->pmsa_nat_bits, &regs->r29, nat);
ms                947 arch/ia64/kernel/mca.c 	copy_reg(&bank[30-16], ms->pmsa_nat_bits, &regs->r30, nat);
ms                948 arch/ia64/kernel/mca.c 	copy_reg(&bank[31-16], ms->pmsa_nat_bits, &regs->r31, nat);
ms                972 arch/ia64/kernel/mca.c 	const pal_min_state_area_t *ms = sos->pal_min_state;
ms                982 arch/ia64/kernel/mca.c 	u64 r12 = ms->pmsa_gr[12-1], r13 = ms->pmsa_gr[13-1];
ms                996 arch/ia64/kernel/mca.c 	regs->cr_ipsr = ms->pmsa_ipsr;
ms               1053 arch/ia64/kernel/mca.c 	if (!mca_recover_range(ms->pmsa_iip)) {
ms               1112 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[4-1], ms->pmsa_nat_bits, &old_sw->r4, &old_unat);
ms               1113 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[5-1], ms->pmsa_nat_bits, &old_sw->r5, &old_unat);
ms               1114 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[6-1], ms->pmsa_nat_bits, &old_sw->r6, &old_unat);
ms               1115 arch/ia64/kernel/mca.c 	copy_reg(&ms->pmsa_gr[7-1], ms->pmsa_nat_bits, &old_sw->r7, &old_unat);
ms               1117 arch/ia64/kernel/mca.c 	old_sw->b1 = ms->pmsa_br1;
ms                 94 arch/mips/kernel/spinlock_test.c 	struct spin_multi_state ms;
ms                 97 arch/mips/kernel/spinlock_test.c 	ms.lock = __RAW_SPIN_LOCK_UNLOCKED("multi_get");
ms                 98 arch/mips/kernel/spinlock_test.c 	ms.loops = 1000000;
ms                100 arch/mips/kernel/spinlock_test.c 	atomic_set(&ms.start_wait, 2);
ms                101 arch/mips/kernel/spinlock_test.c 	atomic_set(&ms.enter_wait, 2);
ms                102 arch/mips/kernel/spinlock_test.c 	atomic_set(&ms.exit_wait, 2);
ms                103 arch/mips/kernel/spinlock_test.c 	t1.state = &ms;
ms                104 arch/mips/kernel/spinlock_test.c 	t2.state = &ms;
ms                492 arch/powerpc/kernel/rtas.c 	unsigned int ms = 0;
ms                495 arch/powerpc/kernel/rtas.c 		ms = 1;
ms                499 arch/powerpc/kernel/rtas.c 		for (ms = 1; order > 0; order--)
ms                500 arch/powerpc/kernel/rtas.c 			ms *= 10;
ms                503 arch/powerpc/kernel/rtas.c 	return ms;
ms                510 arch/powerpc/kernel/rtas.c 	unsigned int ms;
ms                513 arch/powerpc/kernel/rtas.c 	ms = rtas_busy_delay_time(status);
ms                514 arch/powerpc/kernel/rtas.c 	if (ms && need_resched())
ms                515 arch/powerpc/kernel/rtas.c 		msleep(ms);
ms                517 arch/powerpc/kernel/rtas.c 	return ms;
ms                117 arch/powerpc/sysdev/xive/spapr.c 	unsigned int ms = 0;
ms                120 arch/powerpc/sysdev/xive/spapr.c 		ms = get_longbusy_msecs(rc);
ms                122 arch/powerpc/sysdev/xive/spapr.c 		ms = 10; /* seems appropriate for XIVE hcalls */
ms                125 arch/powerpc/sysdev/xive/spapr.c 	return ms;
ms                130 arch/powerpc/sysdev/xive/spapr.c 	unsigned int ms;
ms                132 arch/powerpc/sysdev/xive/spapr.c 	ms = plpar_busy_delay_time(rc);
ms                133 arch/powerpc/sysdev/xive/spapr.c 	if (ms)
ms                134 arch/powerpc/sysdev/xive/spapr.c 		mdelay(ms);
ms                136 arch/powerpc/sysdev/xive/spapr.c 	return ms;
ms                 58 arch/s390/include/asm/nmi.h 		u64 ms :  1; /* 21 psw mask and key validity */
ms                 29 arch/s390/kernel/machine_kexec_file.c 	struct module_signature *ms;
ms                 44 arch/s390/kernel/machine_kexec_file.c 	ms = (void *)kernel + kernel_len - sizeof(*ms);
ms                 45 arch/s390/kernel/machine_kexec_file.c 	kernel_len -= sizeof(*ms);
ms                 47 arch/s390/kernel/machine_kexec_file.c 	sig_len = be32_to_cpu(ms->sig_len);
ms                 52 arch/s390/kernel/machine_kexec_file.c 	if (ms->id_type != PKEY_ID_PKCS7)
ms                 55 arch/s390/kernel/machine_kexec_file.c 	if (ms->algo != 0 ||
ms                 56 arch/s390/kernel/machine_kexec_file.c 	    ms->hash != 0 ||
ms                 57 arch/s390/kernel/machine_kexec_file.c 	    ms->signer_len != 0 ||
ms                 58 arch/s390/kernel/machine_kexec_file.c 	    ms->key_id_len != 0 ||
ms                 59 arch/s390/kernel/machine_kexec_file.c 	    ms->__pad[0] != 0 ||
ms                 60 arch/s390/kernel/machine_kexec_file.c 	    ms->__pad[1] != 0 ||
ms                 61 arch/s390/kernel/machine_kexec_file.c 	    ms->__pad[2] != 0) {
ms                292 arch/s390/kernel/nmi.c 	if (!mci.ms || !mci.pm || !mci.ia)
ms               1002 arch/s390/kvm/kvm-s390.c 	struct kvm_memory_slot *ms;
ms               1020 arch/s390/kvm/kvm-s390.c 		ms = slots->memslots + slotnr;
ms               1021 arch/s390/kvm/kvm-s390.c 		if (!ms->dirty_bitmap)
ms               1029 arch/s390/kvm/kvm-s390.c 		memset(kvm_second_dirty_bitmap(ms), 0xff, kvm_dirty_bitmap_bytes(ms));
ms               1030 arch/s390/kvm/kvm-s390.c 		ram_pages += ms->npages;
ms               1973 arch/s390/kvm/kvm-s390.c 	struct kvm_memory_slot *ms = slots->memslots + slotidx;
ms               1974 arch/s390/kvm/kvm-s390.c 	unsigned long ofs = cur_gfn - ms->base_gfn;
ms               1976 arch/s390/kvm/kvm-s390.c 	if (ms->base_gfn + ms->npages <= cur_gfn) {
ms               1982 arch/s390/kvm/kvm-s390.c 		ms = slots->memslots + slotidx;
ms               1985 arch/s390/kvm/kvm-s390.c 	ofs = find_next_bit(kvm_second_dirty_bitmap(ms), ms->npages, ofs);
ms               1986 arch/s390/kvm/kvm-s390.c 	while ((slotidx > 0) && (ofs >= ms->npages)) {
ms               1988 arch/s390/kvm/kvm-s390.c 		ms = slots->memslots + slotidx;
ms               1989 arch/s390/kvm/kvm-s390.c 		ofs = find_next_bit(kvm_second_dirty_bitmap(ms), ms->npages, 0);
ms               1991 arch/s390/kvm/kvm-s390.c 	return ms->base_gfn + ofs;
ms               1999 arch/s390/kvm/kvm-s390.c 	struct kvm_memory_slot *ms;
ms               2002 arch/s390/kvm/kvm-s390.c 	ms = gfn_to_memslot(kvm, cur_gfn);
ms               2005 arch/s390/kvm/kvm-s390.c 	if (!ms)
ms               2015 arch/s390/kvm/kvm-s390.c 		if (test_and_clear_bit(cur_gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms)))
ms               2033 arch/s390/kvm/kvm-s390.c 		if (cur_gfn - ms->base_gfn >= ms->npages) {
ms               2034 arch/s390/kvm/kvm-s390.c 			ms = gfn_to_memslot(kvm, cur_gfn);
ms               2035 arch/s390/kvm/kvm-s390.c 			if (!ms)
ms               1174 arch/s390/kvm/priv.c 		struct kvm_memory_slot *ms = gfn_to_memslot(vcpu->kvm, gfn);
ms               1177 arch/s390/kvm/priv.c 		if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms)))
ms                485 arch/x86/events/perf_event.h #define EVENT_EXTRA_REG(e, ms, m, vm, i) {	\
ms                487 arch/x86/events/perf_event.h 	.msr = (ms),			\
ms               1126 arch/x86/kernel/cpu/amd.c 	u32 ms;
ms               1143 arch/x86/kernel/cpu/amd.c 	ms = (cpu->x86_model << 4) | cpu->x86_stepping;
ms               1146 arch/x86/kernel/cpu/amd.c 		    (ms >= AMD_MODEL_RANGE_START(range)) &&
ms               1147 arch/x86/kernel/cpu/amd.c 		    (ms <= AMD_MODEL_RANGE_END(range)))
ms                381 arch/x86/kernel/tsc.c static unsigned long pit_calibrate_tsc(u32 latch, unsigned long ms, int loopmin)
ms                443 arch/x86/kernel/tsc.c 	do_div(delta, ms);
ms                713 arch/x86/kernel/tsc.c 	unsigned long flags, latch, ms;
ms                743 arch/x86/kernel/tsc.c 	ms = CAL_MS;
ms                757 arch/x86/kernel/tsc.c 		tsc_pit_khz = pit_calibrate_tsc(latch, ms, loopmin);
ms                804 arch/x86/kernel/tsc.c 			ms = CAL2_MS;
ms                607 drivers/acpi/osl.c void acpi_os_sleep(u64 ms)
ms                609 drivers/acpi/osl.c 	msleep(ms);
ms                100 drivers/auxdisplay/charlcd.c static void long_sleep(int ms)
ms                102 drivers/auxdisplay/charlcd.c 	schedule_timeout_interruptible(msecs_to_jiffies(ms));
ms                 86 drivers/bcma/driver_chipcommon.c 						 u32 ms)
ms                 91 drivers/bcma/driver_chipcommon.c 	ticks = bcma_chipco_watchdog_timer_set(cc, cc->ticks_per_ms * ms);
ms                339 drivers/block/amiflop.c static void ms_delay(int ms)
ms                344 drivers/block/amiflop.c 	if (ms > 0) {
ms                346 drivers/block/amiflop.c 		ticks = MS_TICKS*ms-1;
ms               1150 drivers/block/drbd/drbd_bitmap.c 		unsigned int ms = jiffies_to_msecs(jiffies - now);
ms               1151 drivers/block/drbd/drbd_bitmap.c 		if (ms > 5) {
ms               1154 drivers/block/drbd/drbd_bitmap.c 				 count, ms);
ms               4358 drivers/block/drbd/drbd_receiver.c 	union drbd_state ms;
ms               4371 drivers/block/drbd/drbd_receiver.c 	ms.i = ps.i;
ms               4373 drivers/block/drbd/drbd_receiver.c 	ms.conn = c_tab[ps.conn];
ms               4374 drivers/block/drbd/drbd_receiver.c 	ms.peer = ps.role;
ms               4375 drivers/block/drbd/drbd_receiver.c 	ms.role = ps.peer;
ms               4376 drivers/block/drbd/drbd_receiver.c 	ms.pdsk = ps.disk;
ms               4377 drivers/block/drbd/drbd_receiver.c 	ms.disk = ps.pdsk;
ms               4378 drivers/block/drbd/drbd_receiver.c 	ms.peer_isp = (ps.aftr_isp | ps.user_isp);
ms               4380 drivers/block/drbd/drbd_receiver.c 	return ms;
ms                665 drivers/block/paride/pcd.c static int pcd_probe(struct pcd_unit *cd, int ms, char *id)
ms                667 drivers/block/paride/pcd.c 	if (ms == -1) {
ms                672 drivers/block/paride/pcd.c 		cd->drive = ms;
ms                651 drivers/block/paride/pf.c 	char *ms[2] = { "master", "slave" };
ms                681 drivers/block/paride/pf.c 	       pf->name, mf, id, ms[pf->drive], pf->lun, dt);
ms                440 drivers/block/paride/pg.c 	char *ms[2] = { "master", "slave" };
ms                455 drivers/block/paride/pg.c 		printk("%s: %s %s, %s\n", dev->name, mf, id, ms[dev->drive]);
ms                531 drivers/block/paride/pt.c 	char *ms[2] = { "master", "slave" };
ms                571 drivers/block/paride/pt.c 	printk("%s: %s %s, %s", tape->name, mf, id, ms[tape->drive]);
ms                425 drivers/clk/rockchip/clk.h #define COMPOSITE(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\
ms                435 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                446 drivers/clk/rockchip/clk.h #define COMPOSITE_DIV_OFFSET(_id, cname, pnames, f, mo, ms, mw,	\
ms                456 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                505 drivers/clk/rockchip/clk.h #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf,	\
ms                515 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                523 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE(_id, cname, pnames, f, mo, ms, mw, mf,	\
ms                533 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                542 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE_DIVTBL(_id, cname, pnames, f, mo, ms,	\
ms                552 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                613 drivers/clk/rockchip/clk.h #define COMPOSITE_DDRCLK(_id, cname, pnames, f, mo, ms, mw,	\
ms                623 drivers/clk/rockchip/clk.h 		.mux_shift      = ms,                           \
ms                754 drivers/clk/rockchip/clk.h #define COMPOSITE_HALFDIV(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\
ms                764 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                775 drivers/clk/rockchip/clk.h #define COMPOSITE_NOGATE_HALFDIV(_id, cname, pnames, f, mo, ms, mw, mf,	\
ms                785 drivers/clk/rockchip/clk.h 		.mux_shift	= ms,				\
ms                198 drivers/clocksource/timer-tegra.c 	u32 ms = readl_relaxed(reg_base + RTC_MILLISECONDS);
ms                201 drivers/clocksource/timer-tegra.c 	return (u64)s * MSEC_PER_SEC + ms;
ms                 87 drivers/cpufreq/pmac32-cpufreq.c static inline void local_delay(unsigned long ms)
ms                 90 drivers/cpufreq/pmac32-cpufreq.c 		mdelay(ms);
ms                 92 drivers/cpufreq/pmac32-cpufreq.c 		msleep(ms);
ms                279 drivers/dma/imx-sdma.c 	u32  ms;
ms                392 drivers/gpu/drm/arm/malidp_crtc.c 		struct malidp_plane_state *ms = to_malidp_plane_state(pstate);
ms                412 drivers/gpu/drm/arm/malidp_crtc.c 			if (ms->rotmem_size > rot_mem_usable)
ms                138 drivers/gpu/drm/arm/malidp_planes.c 	struct malidp_plane_state *ms = to_malidp_plane_state(state);
ms                140 drivers/gpu/drm/arm/malidp_planes.c 	drm_printf(p, "\trotmem_size=%u\n", ms->rotmem_size);
ms                141 drivers/gpu/drm/arm/malidp_planes.c 	drm_printf(p, "\tformat_id=%u\n", ms->format);
ms                142 drivers/gpu/drm/arm/malidp_planes.c 	drm_printf(p, "\tn_planes=%u\n", ms->n_planes);
ms                144 drivers/gpu/drm/arm/malidp_planes.c 		   prefetch_mode_names[ms->mmu_prefetch_mode]);
ms                145 drivers/gpu/drm/arm/malidp_planes.c 	drm_printf(p, "\tmmu_prefetch_pgsize=%d\n", ms->mmu_prefetch_pgsize);
ms                332 drivers/gpu/drm/arm/malidp_planes.c static bool malidp_check_pages_threshold(struct malidp_plane_state *ms,
ms                337 drivers/gpu/drm/arm/malidp_planes.c 	for (i = 0; i < ms->n_planes; i++) {
ms                343 drivers/gpu/drm/arm/malidp_planes.c 		obj = drm_gem_fb_get_obj(ms->base.fb, i);
ms                436 drivers/gpu/drm/arm/malidp_planes.c 		(struct malidp_plane_state *ms,	u32 *pgsize_bitmap)
ms                446 drivers/gpu/drm/arm/malidp_planes.c 		if (malidp_check_pages_threshold(ms, largest_pgsize)) {
ms                457 drivers/gpu/drm/arm/malidp_planes.c 	if (malidp_partial_prefetch_supported(ms->base.fb->format->format,
ms                458 drivers/gpu/drm/arm/malidp_planes.c 					      ms->base.fb->modifier,
ms                459 drivers/gpu/drm/arm/malidp_planes.c 					      ms->base.rotation)) {
ms                493 drivers/gpu/drm/arm/malidp_planes.c 					struct malidp_plane_state *ms)
ms                499 drivers/gpu/drm/arm/malidp_planes.c 	ms->mmu_prefetch_pgsize = malidp_get_pgsize_bitmap(mp);
ms                500 drivers/gpu/drm/arm/malidp_planes.c 	ms->mmu_prefetch_mode  =
ms                501 drivers/gpu/drm/arm/malidp_planes.c 		malidp_mmu_prefetch_select_mode(ms, &ms->mmu_prefetch_pgsize);
ms                508 drivers/gpu/drm/arm/malidp_planes.c 	struct malidp_plane_state *ms = to_malidp_plane_state(state);
ms                520 drivers/gpu/drm/arm/malidp_planes.c 	ms->format = malidp_hw_get_format_id(&mp->hwdev->hw->map,
ms                523 drivers/gpu/drm/arm/malidp_planes.c 	if (ms->format == MALIDP_INVALID_FORMAT_ID)
ms                526 drivers/gpu/drm/arm/malidp_planes.c 	ms->n_planes = fb->format->num_planes;
ms                527 drivers/gpu/drm/arm/malidp_planes.c 	for (i = 0; i < ms->n_planes; i++) {
ms                560 drivers/gpu/drm/arm/malidp_planes.c 	if (ms->n_planes == 3 &&
ms                590 drivers/gpu/drm/arm/malidp_planes.c 	ms->rotmem_size = 0;
ms                601 drivers/gpu/drm/arm/malidp_planes.c 		ms->rotmem_size = val;
ms                610 drivers/gpu/drm/arm/malidp_planes.c 	malidp_de_prefetch_settings(mp, ms);
ms                697 drivers/gpu/drm/arm/malidp_planes.c 				      struct malidp_plane_state *ms)
ms                705 drivers/gpu/drm/arm/malidp_planes.c 	mmu_ctrl = malidp_calc_mmu_control_value(ms->mmu_prefetch_mode,
ms                707 drivers/gpu/drm/arm/malidp_planes.c 						 ms->n_planes,
ms                708 drivers/gpu/drm/arm/malidp_planes.c 						 ms->mmu_prefetch_pgsize);
ms                795 drivers/gpu/drm/arm/malidp_planes.c 	struct malidp_plane_state *ms = to_malidp_plane_state(plane->state);
ms                822 drivers/gpu/drm/arm/malidp_planes.c 	val = (val & ~LAYER_FORMAT_MASK) | ms->format;
ms                825 drivers/gpu/drm/arm/malidp_planes.c 	for (i = 0; i < ms->n_planes; i++)
ms                828 drivers/gpu/drm/arm/malidp_planes.c 	malidp_de_set_mmu_control(mp, ms);
ms                830 drivers/gpu/drm/arm/malidp_planes.c 	malidp_de_set_plane_pitches(mp, ms->n_planes,
ms                284 drivers/gpu/drm/i915/selftests/i915_buddy.c 	u64 s, ms;
ms                291 drivers/gpu/drm/i915/selftests/i915_buddy.c 	ms = BIT_ULL(12 + (prandom_u32_state(&prng) % ilog2(s >> 12)));
ms                292 drivers/gpu/drm/i915/selftests/i915_buddy.c 	s = max(s & -ms, ms);
ms                294 drivers/gpu/drm/i915/selftests/i915_buddy.c 	*chunk_size = ms;
ms                241 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c static int msm_hdmi_hdcp_msleep(struct hdmi_hdcp_ctrl *hdcp_ctrl, u32 ms, u32 ev)
ms                247 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c 		msecs_to_jiffies(ms));
ms                126 drivers/gpu/drm/nouveau/nvkm/falcon/base.c nvkm_falcon_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
ms                128 drivers/gpu/drm/nouveau/nvkm/falcon/base.c 	return falcon->func->wait_for_halt(falcon, ms);
ms                292 drivers/gpu/drm/nouveau/nvkm/falcon/v1.c nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
ms                297 drivers/gpu/drm/nouveau/nvkm/falcon/v1.c 	ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
ms                 87 drivers/gpu/drm/via/via_video.c 			    (fx->ms / 10) * (HZ / 100), *lock != fx->val);
ms                590 drivers/gpu/ipu-v3/ipu-common.c int ipu_idmac_wait_busy(struct ipuv3_channel *channel, int ms)
ms                595 drivers/gpu/ipu-v3/ipu-common.c 	timeout = jiffies + msecs_to_jiffies(ms);
ms                 62 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                 63 drivers/hid/hid-microsoft.c 	unsigned long quirks = ms->quirks;
ms                189 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                190 drivers/hid/hid-microsoft.c 	unsigned long quirks = ms->quirks;
ms                216 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                217 drivers/hid/hid-microsoft.c 	unsigned long quirks = ms->quirks;
ms                228 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                229 drivers/hid/hid-microsoft.c 	unsigned long quirks = ms->quirks;
ms                286 drivers/hid/hid-microsoft.c 	struct ms_data *ms = container_of(work, struct ms_data, ff_worker);
ms                287 drivers/hid/hid-microsoft.c 	struct hid_device *hdev = ms->hdev;
ms                288 drivers/hid/hid-microsoft.c 	struct xb1s_ff_report *r = ms->output_report_dmabuf;
ms                302 drivers/hid/hid-microsoft.c 	r->magnitude[MAGNITUDE_STRONG] = ms->strong; /* left actuator */
ms                303 drivers/hid/hid-microsoft.c 	r->magnitude[MAGNITUDE_WEAK] = ms->weak;     /* right actuator */
ms                314 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hid);
ms                322 drivers/hid/hid-microsoft.c 	ms->strong = ((u32) effect->u.rumble.strong_magnitude * 100) / U16_MAX;
ms                323 drivers/hid/hid-microsoft.c 	ms->weak = ((u32) effect->u.rumble.weak_magnitude * 100) / U16_MAX;
ms                325 drivers/hid/hid-microsoft.c 	schedule_work(&ms->ff_worker);
ms                333 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                342 drivers/hid/hid-microsoft.c 	if (!(ms->quirks & MS_QUIRK_FF))
ms                345 drivers/hid/hid-microsoft.c 	ms->hdev = hdev;
ms                346 drivers/hid/hid-microsoft.c 	INIT_WORK(&ms->ff_worker, ms_ff_worker);
ms                348 drivers/hid/hid-microsoft.c 	ms->output_report_dmabuf = devm_kzalloc(&hdev->dev,
ms                351 drivers/hid/hid-microsoft.c 	if (ms->output_report_dmabuf == NULL)
ms                360 drivers/hid/hid-microsoft.c 	struct ms_data *ms = hid_get_drvdata(hdev);
ms                362 drivers/hid/hid-microsoft.c 	if (!(ms->quirks & MS_QUIRK_FF))
ms                365 drivers/hid/hid-microsoft.c 	cancel_work_sync(&ms->ff_worker);
ms                371 drivers/hid/hid-microsoft.c 	struct ms_data *ms;
ms                374 drivers/hid/hid-microsoft.c 	ms = devm_kzalloc(&hdev->dev, sizeof(*ms), GFP_KERNEL);
ms                375 drivers/hid/hid-microsoft.c 	if (ms == NULL)
ms                378 drivers/hid/hid-microsoft.c 	ms->quirks = quirks;
ms                380 drivers/hid/hid-microsoft.c 	hid_set_drvdata(hdev, ms);
ms                614 drivers/hwmon/applesmc.c 	int ms, ret;
ms                616 drivers/hwmon/applesmc.c 	for (ms = 0; ms < INIT_TIMEOUT_MSECS; ms += INIT_WAIT_MSECS) {
ms                619 drivers/hwmon/applesmc.c 			if (ms)
ms                620 drivers/hwmon/applesmc.c 				pr_info("init_smcreg() took %d ms\n", ms);
ms                182 drivers/hwtracing/stm/p_sys-t.c 	unsigned int ms;
ms                186 drivers/hwtracing/stm/p_sys-t.c 	ret = kstrtouint(page, 10, &ms);
ms                190 drivers/hwtracing/stm/p_sys-t.c 		pn->ts_interval = msecs_to_jiffies(ms);
ms                213 drivers/hwtracing/stm/p_sys-t.c 	unsigned int ms;
ms                217 drivers/hwtracing/stm/p_sys-t.c 	ret = kstrtouint(page, 10, &ms);
ms                221 drivers/hwtracing/stm/p_sys-t.c 		pn->clocksync_interval = msecs_to_jiffies(ms);
ms                265 drivers/i2c/busses/i2c-diolan-u2c.c static int diolan_set_clock_synch_timeout(struct i2c_diolan_u2c *dev, int ms)
ms                267 drivers/i2c/busses/i2c-diolan-u2c.c 	int to_val = ms * 10;
ms                 71 drivers/input/serio/i8042-sparcio.h 			struct platform_device *ms = of_find_device_by_node(dp);
ms                 72 drivers/input/serio/i8042-sparcio.h 			unsigned int irq = ms->archdata.irqs[0];
ms                 86 drivers/input/touchscreen/auo-pixcir-ts.c #define AUO_PIXCIR_POWER_IDLE_TIME(ms)	((ms & 0xf) << 4)
ms                 25 drivers/leds/leds-asic3.c #define MS_TO_CLK(ms)	DIV_ROUND_CLOSEST(((ms)*1024), 32000)
ms                153 drivers/md/bcache/io.c 			int ms = us / 1024;
ms                157 drivers/md/bcache/io.c 			ms = min(ms, CONGESTED_MAX + congested);
ms                158 drivers/md/bcache/io.c 			atomic_sub(ms, &c->congested);
ms                 81 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_gc,	sec, ms);
ms                 83 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_sort,	ms,  us);
ms                 84 drivers/md/bcache/sysfs.c sysfs_time_stats_attribute(btree_read,	ms,  us);
ms                720 drivers/md/bcache/sysfs.c 	sysfs_print_time_stats(&c->btree_gc_time,	btree_gc, sec, ms);
ms                722 drivers/md/bcache/sysfs.c 	sysfs_print_time_stats(&c->sort.time,		btree_sort, ms, us);
ms                723 drivers/md/bcache/sysfs.c 	sysfs_print_time_stats(&c->btree_read_time,	btree_read, ms, us);
ms                952 drivers/md/bcache/sysfs.c 	sysfs_time_stats_attribute_list(btree_gc, sec, ms)
ms                954 drivers/md/bcache/sysfs.c 	sysfs_time_stats_attribute_list(btree_sort, ms, us)
ms                955 drivers/md/bcache/sysfs.c 	sysfs_time_stats_attribute_list(btree_read, ms, us)
ms                416 drivers/md/dm-integrity.c 	__u64 ms;
ms                419 drivers/md/dm-integrity.c 	ms = area << ic->sb->log2_interleave_sectors;
ms                421 drivers/md/dm-integrity.c 		ms += area << ic->log2_metadata_run;
ms                423 drivers/md/dm-integrity.c 		ms += area * ic->metadata_run;
ms                424 drivers/md/dm-integrity.c 	ms >>= ic->log2_buffer_sectors;
ms                429 drivers/md/dm-integrity.c 		ms += offset >> (SECTOR_SHIFT + ic->log2_buffer_sectors - ic->log2_tag_size);
ms                432 drivers/md/dm-integrity.c 		ms += (__u64)offset * ic->tag_size >> (SECTOR_SHIFT + ic->log2_buffer_sectors);
ms                436 drivers/md/dm-integrity.c 	return ms;
ms                 46 drivers/md/dm-raid1.c 	struct mirror_set *ms;
ms                 94 drivers/md/dm-raid1.c 	struct mirror_set *ms = context;
ms                 96 drivers/md/dm-raid1.c 	queue_work(ms->kmirrord_wq, &ms->kmirrord_work);
ms                101 drivers/md/dm-raid1.c 	struct mirror_set *ms = from_timer(ms, t, timer);
ms                103 drivers/md/dm-raid1.c 	clear_bit(0, &ms->timer_pending);
ms                104 drivers/md/dm-raid1.c 	wakeup_mirrord(ms);
ms                107 drivers/md/dm-raid1.c static void delayed_wake(struct mirror_set *ms)
ms                109 drivers/md/dm-raid1.c 	if (test_and_set_bit(0, &ms->timer_pending))
ms                112 drivers/md/dm-raid1.c 	ms->timer.expires = jiffies + HZ / 5;
ms                113 drivers/md/dm-raid1.c 	add_timer(&ms->timer);
ms                121 drivers/md/dm-raid1.c static void queue_bio(struct mirror_set *ms, struct bio *bio, int rw)
ms                127 drivers/md/dm-raid1.c 	bl = (rw == WRITE) ? &ms->writes : &ms->reads;
ms                128 drivers/md/dm-raid1.c 	spin_lock_irqsave(&ms->lock, flags);
ms                131 drivers/md/dm-raid1.c 	spin_unlock_irqrestore(&ms->lock, flags);
ms                134 drivers/md/dm-raid1.c 		wakeup_mirrord(ms);
ms                139 drivers/md/dm-raid1.c 	struct mirror_set *ms = context;
ms                143 drivers/md/dm-raid1.c 		queue_bio(ms, bio, WRITE);
ms                173 drivers/md/dm-raid1.c static struct mirror *get_default_mirror(struct mirror_set *ms)
ms                175 drivers/md/dm-raid1.c 	return &ms->mirror[atomic_read(&ms->default_mirror)];
ms                180 drivers/md/dm-raid1.c 	struct mirror_set *ms = m->ms;
ms                181 drivers/md/dm-raid1.c 	struct mirror *m0 = &(ms->mirror[0]);
ms                183 drivers/md/dm-raid1.c 	atomic_set(&ms->default_mirror, m - m0);
ms                186 drivers/md/dm-raid1.c static struct mirror *get_valid_mirror(struct mirror_set *ms)
ms                190 drivers/md/dm-raid1.c 	for (m = ms->mirror; m < ms->mirror + ms->nr_mirrors; m++)
ms                213 drivers/md/dm-raid1.c 	struct mirror_set *ms = m->ms;
ms                216 drivers/md/dm-raid1.c 	ms->leg_failure = 1;
ms                228 drivers/md/dm-raid1.c 	if (!errors_handled(ms))
ms                231 drivers/md/dm-raid1.c 	if (m != get_default_mirror(ms))
ms                234 drivers/md/dm-raid1.c 	if (!ms->in_sync && !keep_log(ms)) {
ms                244 drivers/md/dm-raid1.c 	new = get_valid_mirror(ms);
ms                251 drivers/md/dm-raid1.c 	schedule_work(&ms->trigger_event);
ms                256 drivers/md/dm-raid1.c 	struct mirror_set *ms = ti->private;
ms                267 drivers/md/dm-raid1.c 		.client = ms->io_client,
ms                270 drivers/md/dm-raid1.c 	for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++) {
ms                277 drivers/md/dm-raid1.c 	dm_io(&io_req, ms->nr_mirrors, io, &error_bits);
ms                279 drivers/md/dm-raid1.c 		for (i = 0; i < ms->nr_mirrors; i++)
ms                281 drivers/md/dm-raid1.c 				fail_mirror(ms->mirror + i,
ms                300 drivers/md/dm-raid1.c 	struct mirror_set *ms = dm_rh_region_context(reg);
ms                306 drivers/md/dm-raid1.c 		fail_mirror(get_default_mirror(ms), DM_RAID1_SYNC_ERROR);
ms                316 drivers/md/dm-raid1.c 		for (m = 0; m < ms->nr_mirrors; m++) {
ms                317 drivers/md/dm-raid1.c 			if (&ms->mirror[m] == get_default_mirror(ms))
ms                320 drivers/md/dm-raid1.c 				fail_mirror(ms->mirror + m,
ms                329 drivers/md/dm-raid1.c static void recover(struct mirror_set *ms, struct dm_region *reg)
ms                336 drivers/md/dm-raid1.c 	sector_t region_size = dm_rh_get_region_size(ms->rh);
ms                339 drivers/md/dm-raid1.c 	m = get_default_mirror(ms);
ms                341 drivers/md/dm-raid1.c 	from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
ms                342 drivers/md/dm-raid1.c 	if (key == (ms->nr_regions - 1)) {
ms                347 drivers/md/dm-raid1.c 		from.count = ms->ti->len & (region_size - 1);
ms                354 drivers/md/dm-raid1.c 	for (i = 0, dest = to; i < ms->nr_mirrors; i++) {
ms                355 drivers/md/dm-raid1.c 		if (&ms->mirror[i] == get_default_mirror(ms))
ms                358 drivers/md/dm-raid1.c 		m = ms->mirror + i;
ms                360 drivers/md/dm-raid1.c 		dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
ms                366 drivers/md/dm-raid1.c 	if (!errors_handled(ms))
ms                369 drivers/md/dm-raid1.c 	dm_kcopyd_copy(ms->kcopyd_client, &from, ms->nr_mirrors - 1, to,
ms                373 drivers/md/dm-raid1.c static void reset_ms_flags(struct mirror_set *ms)
ms                377 drivers/md/dm-raid1.c 	ms->leg_failure = 0;
ms                378 drivers/md/dm-raid1.c 	for (m = 0; m < ms->nr_mirrors; m++) {
ms                379 drivers/md/dm-raid1.c 		atomic_set(&(ms->mirror[m].error_count), 0);
ms                380 drivers/md/dm-raid1.c 		ms->mirror[m].error_type = 0;
ms                384 drivers/md/dm-raid1.c static void do_recovery(struct mirror_set *ms)
ms                387 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms                392 drivers/md/dm-raid1.c 	dm_rh_recovery_prepare(ms->rh);
ms                397 drivers/md/dm-raid1.c 	while ((reg = dm_rh_recovery_start(ms->rh)))
ms                398 drivers/md/dm-raid1.c 		recover(ms, reg);
ms                403 drivers/md/dm-raid1.c 	if (!ms->in_sync &&
ms                404 drivers/md/dm-raid1.c 	    (log->type->get_sync_count(log) == ms->nr_regions)) {
ms                406 drivers/md/dm-raid1.c 		dm_table_event(ms->ti->table);
ms                407 drivers/md/dm-raid1.c 		ms->in_sync = 1;
ms                408 drivers/md/dm-raid1.c 		reset_ms_flags(ms);
ms                415 drivers/md/dm-raid1.c static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector)
ms                417 drivers/md/dm-raid1.c 	struct mirror *m = get_default_mirror(ms);
ms                423 drivers/md/dm-raid1.c 		if (m-- == ms->mirror)
ms                424 drivers/md/dm-raid1.c 			m += ms->nr_mirrors;
ms                425 drivers/md/dm-raid1.c 	} while (m != get_default_mirror(ms));
ms                432 drivers/md/dm-raid1.c 	struct mirror *default_mirror = get_default_mirror(m->ms);
ms                437 drivers/md/dm-raid1.c static int mirror_available(struct mirror_set *ms, struct bio *bio)
ms                439 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms                440 drivers/md/dm-raid1.c 	region_t region = dm_rh_bio_to_region(ms->rh, bio);
ms                443 drivers/md/dm-raid1.c 		return choose_mirror(ms,  bio->bi_iter.bi_sector) ? 1 : 0;
ms                455 drivers/md/dm-raid1.c 	return m->offset + dm_target_offset(m->ms->ti, bio->bi_iter.bi_sector);
ms                472 drivers/md/dm-raid1.c static void hold_bio(struct mirror_set *ms, struct bio *bio)
ms                478 drivers/md/dm-raid1.c 	spin_lock_irq(&ms->lock);
ms                480 drivers/md/dm-raid1.c 	if (atomic_read(&ms->suspend)) {
ms                481 drivers/md/dm-raid1.c 		spin_unlock_irq(&ms->lock);
ms                486 drivers/md/dm-raid1.c 		if (dm_noflush_suspending(ms->ti))
ms                498 drivers/md/dm-raid1.c 	bio_list_add(&ms->holds, bio);
ms                499 drivers/md/dm-raid1.c 	spin_unlock_irq(&ms->lock);
ms                520 drivers/md/dm-raid1.c 	if (likely(default_ok(m)) || mirror_available(m->ms, bio)) {
ms                524 drivers/md/dm-raid1.c 		queue_bio(m->ms, bio, bio_data_dir(bio));
ms                544 drivers/md/dm-raid1.c 		.client = m->ms->io_client,
ms                552 drivers/md/dm-raid1.c static inline int region_in_sync(struct mirror_set *ms, region_t region,
ms                555 drivers/md/dm-raid1.c 	int state = dm_rh_get_state(ms->rh, region, may_block);
ms                559 drivers/md/dm-raid1.c static void do_reads(struct mirror_set *ms, struct bio_list *reads)
ms                566 drivers/md/dm-raid1.c 		region = dm_rh_bio_to_region(ms->rh, bio);
ms                567 drivers/md/dm-raid1.c 		m = get_default_mirror(ms);
ms                572 drivers/md/dm-raid1.c 		if (likely(region_in_sync(ms, region, 1)))
ms                573 drivers/md/dm-raid1.c 			m = choose_mirror(ms, bio->bi_iter.bi_sector);
ms                600 drivers/md/dm-raid1.c 	struct mirror_set *ms;
ms                604 drivers/md/dm-raid1.c 	ms = bio_get_m(bio)->ms;
ms                628 drivers/md/dm-raid1.c 	for (i = 0; i < ms->nr_mirrors; i++)
ms                630 drivers/md/dm-raid1.c 			fail_mirror(ms->mirror + i, DM_RAID1_WRITE_ERROR);
ms                637 drivers/md/dm-raid1.c 	spin_lock_irqsave(&ms->lock, flags);
ms                638 drivers/md/dm-raid1.c 	if (!ms->failures.head)
ms                640 drivers/md/dm-raid1.c 	bio_list_add(&ms->failures, bio);
ms                641 drivers/md/dm-raid1.c 	spin_unlock_irqrestore(&ms->lock, flags);
ms                643 drivers/md/dm-raid1.c 		wakeup_mirrord(ms);
ms                646 drivers/md/dm-raid1.c static void do_write(struct mirror_set *ms, struct bio *bio)
ms                658 drivers/md/dm-raid1.c 		.client = ms->io_client,
ms                667 drivers/md/dm-raid1.c 	for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++)
ms                674 drivers/md/dm-raid1.c 	bio_set_m(bio, get_default_mirror(ms));
ms                676 drivers/md/dm-raid1.c 	BUG_ON(dm_io(&io_req, ms->nr_mirrors, io, NULL));
ms                679 drivers/md/dm-raid1.c static void do_writes(struct mirror_set *ms, struct bio_list *writes)
ms                685 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms                706 drivers/md/dm-raid1.c 		region = dm_rh_bio_to_region(ms->rh, bio);
ms                714 drivers/md/dm-raid1.c 		state = dm_rh_get_state(ms->rh, region, 1);
ms                738 drivers/md/dm-raid1.c 		spin_lock_irq(&ms->lock);
ms                739 drivers/md/dm-raid1.c 		bio_list_merge(&ms->writes, &requeue);
ms                740 drivers/md/dm-raid1.c 		spin_unlock_irq(&ms->lock);
ms                741 drivers/md/dm-raid1.c 		delayed_wake(ms);
ms                749 drivers/md/dm-raid1.c 	dm_rh_inc_pending(ms->rh, &sync);
ms                750 drivers/md/dm-raid1.c 	dm_rh_inc_pending(ms->rh, &nosync);
ms                757 drivers/md/dm-raid1.c 	ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure;
ms                762 drivers/md/dm-raid1.c 	if (unlikely(ms->log_failure) && errors_handled(ms)) {
ms                763 drivers/md/dm-raid1.c 		spin_lock_irq(&ms->lock);
ms                764 drivers/md/dm-raid1.c 		bio_list_merge(&ms->failures, &sync);
ms                765 drivers/md/dm-raid1.c 		spin_unlock_irq(&ms->lock);
ms                766 drivers/md/dm-raid1.c 		wakeup_mirrord(ms);
ms                769 drivers/md/dm-raid1.c 			do_write(ms, bio);
ms                772 drivers/md/dm-raid1.c 		dm_rh_delay(ms->rh, bio);
ms                775 drivers/md/dm-raid1.c 		if (unlikely(ms->leg_failure) && errors_handled(ms) && !keep_log(ms)) {
ms                776 drivers/md/dm-raid1.c 			spin_lock_irq(&ms->lock);
ms                777 drivers/md/dm-raid1.c 			bio_list_add(&ms->failures, bio);
ms                778 drivers/md/dm-raid1.c 			spin_unlock_irq(&ms->lock);
ms                779 drivers/md/dm-raid1.c 			wakeup_mirrord(ms);
ms                781 drivers/md/dm-raid1.c 			map_bio(get_default_mirror(ms), bio);
ms                787 drivers/md/dm-raid1.c static void do_failures(struct mirror_set *ms, struct bio_list *failures)
ms                812 drivers/md/dm-raid1.c 		if (!ms->log_failure) {
ms                813 drivers/md/dm-raid1.c 			ms->in_sync = 0;
ms                814 drivers/md/dm-raid1.c 			dm_rh_mark_nosync(ms->rh, bio);
ms                830 drivers/md/dm-raid1.c 		if (unlikely(!get_valid_mirror(ms) || (keep_log(ms) && ms->log_failure)))
ms                832 drivers/md/dm-raid1.c 		else if (errors_handled(ms) && !keep_log(ms))
ms                833 drivers/md/dm-raid1.c 			hold_bio(ms, bio);
ms                841 drivers/md/dm-raid1.c 	struct mirror_set *ms =
ms                844 drivers/md/dm-raid1.c 	dm_table_event(ms->ti->table);
ms                852 drivers/md/dm-raid1.c 	struct mirror_set *ms = container_of(work, struct mirror_set,
ms                857 drivers/md/dm-raid1.c 	spin_lock_irqsave(&ms->lock, flags);
ms                858 drivers/md/dm-raid1.c 	reads = ms->reads;
ms                859 drivers/md/dm-raid1.c 	writes = ms->writes;
ms                860 drivers/md/dm-raid1.c 	failures = ms->failures;
ms                861 drivers/md/dm-raid1.c 	bio_list_init(&ms->reads);
ms                862 drivers/md/dm-raid1.c 	bio_list_init(&ms->writes);
ms                863 drivers/md/dm-raid1.c 	bio_list_init(&ms->failures);
ms                864 drivers/md/dm-raid1.c 	spin_unlock_irqrestore(&ms->lock, flags);
ms                866 drivers/md/dm-raid1.c 	dm_rh_update_states(ms->rh, errors_handled(ms));
ms                867 drivers/md/dm-raid1.c 	do_recovery(ms);
ms                868 drivers/md/dm-raid1.c 	do_reads(ms, &reads);
ms                869 drivers/md/dm-raid1.c 	do_writes(ms, &writes);
ms                870 drivers/md/dm-raid1.c 	do_failures(ms, &failures);
ms                881 drivers/md/dm-raid1.c 	struct mirror_set *ms =
ms                882 drivers/md/dm-raid1.c 		kzalloc(struct_size(ms, mirror, nr_mirrors), GFP_KERNEL);
ms                884 drivers/md/dm-raid1.c 	if (!ms) {
ms                889 drivers/md/dm-raid1.c 	spin_lock_init(&ms->lock);
ms                890 drivers/md/dm-raid1.c 	bio_list_init(&ms->reads);
ms                891 drivers/md/dm-raid1.c 	bio_list_init(&ms->writes);
ms                892 drivers/md/dm-raid1.c 	bio_list_init(&ms->failures);
ms                893 drivers/md/dm-raid1.c 	bio_list_init(&ms->holds);
ms                895 drivers/md/dm-raid1.c 	ms->ti = ti;
ms                896 drivers/md/dm-raid1.c 	ms->nr_mirrors = nr_mirrors;
ms                897 drivers/md/dm-raid1.c 	ms->nr_regions = dm_sector_div_up(ti->len, region_size);
ms                898 drivers/md/dm-raid1.c 	ms->in_sync = 0;
ms                899 drivers/md/dm-raid1.c 	ms->log_failure = 0;
ms                900 drivers/md/dm-raid1.c 	ms->leg_failure = 0;
ms                901 drivers/md/dm-raid1.c 	atomic_set(&ms->suspend, 0);
ms                902 drivers/md/dm-raid1.c 	atomic_set(&ms->default_mirror, DEFAULT_MIRROR);
ms                904 drivers/md/dm-raid1.c 	ms->io_client = dm_io_client_create();
ms                905 drivers/md/dm-raid1.c 	if (IS_ERR(ms->io_client)) {
ms                907 drivers/md/dm-raid1.c 		kfree(ms);
ms                911 drivers/md/dm-raid1.c 	ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord,
ms                913 drivers/md/dm-raid1.c 				       ms->ti->begin, MAX_RECOVERY,
ms                914 drivers/md/dm-raid1.c 				       dl, region_size, ms->nr_regions);
ms                915 drivers/md/dm-raid1.c 	if (IS_ERR(ms->rh)) {
ms                917 drivers/md/dm-raid1.c 		dm_io_client_destroy(ms->io_client);
ms                918 drivers/md/dm-raid1.c 		kfree(ms);
ms                922 drivers/md/dm-raid1.c 	return ms;
ms                925 drivers/md/dm-raid1.c static void free_context(struct mirror_set *ms, struct dm_target *ti,
ms                929 drivers/md/dm-raid1.c 		dm_put_device(ti, ms->mirror[m].dev);
ms                931 drivers/md/dm-raid1.c 	dm_io_client_destroy(ms->io_client);
ms                932 drivers/md/dm-raid1.c 	dm_region_hash_destroy(ms->rh);
ms                933 drivers/md/dm-raid1.c 	kfree(ms);
ms                936 drivers/md/dm-raid1.c static int get_mirror(struct mirror_set *ms, struct dm_target *ti,
ms                950 drivers/md/dm-raid1.c 			    &ms->mirror[mirror].dev);
ms                956 drivers/md/dm-raid1.c 	ms->mirror[mirror].ms = ms;
ms                957 drivers/md/dm-raid1.c 	atomic_set(&(ms->mirror[mirror].error_count), 0);
ms                958 drivers/md/dm-raid1.c 	ms->mirror[mirror].error_type = 0;
ms                959 drivers/md/dm-raid1.c 	ms->mirror[mirror].offset = offset;
ms               1002 drivers/md/dm-raid1.c static int parse_features(struct mirror_set *ms, unsigned argc, char **argv,
ms               1006 drivers/md/dm-raid1.c 	struct dm_target *ti = ms->ti;
ms               1031 drivers/md/dm-raid1.c 			ms->features |= DM_RAID1_HANDLE_ERRORS;
ms               1033 drivers/md/dm-raid1.c 			ms->features |= DM_RAID1_KEEP_LOG;
ms               1043 drivers/md/dm-raid1.c 	if (!errors_handled(ms) && keep_log(ms)) {
ms               1067 drivers/md/dm-raid1.c 	struct mirror_set *ms;
ms               1093 drivers/md/dm-raid1.c 	ms = alloc_context(nr_mirrors, dl->type->get_region_size(dl), ti, dl);
ms               1094 drivers/md/dm-raid1.c 	if (!ms) {
ms               1101 drivers/md/dm-raid1.c 		r = get_mirror(ms, ti, m, argv);
ms               1103 drivers/md/dm-raid1.c 			free_context(ms, ti, m);
ms               1110 drivers/md/dm-raid1.c 	ti->private = ms;
ms               1112 drivers/md/dm-raid1.c 	r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh));
ms               1120 drivers/md/dm-raid1.c 	ms->kmirrord_wq = alloc_workqueue("kmirrord", WQ_MEM_RECLAIM, 0);
ms               1121 drivers/md/dm-raid1.c 	if (!ms->kmirrord_wq) {
ms               1126 drivers/md/dm-raid1.c 	INIT_WORK(&ms->kmirrord_work, do_mirror);
ms               1127 drivers/md/dm-raid1.c 	timer_setup(&ms->timer, delayed_wake_fn, 0);
ms               1128 drivers/md/dm-raid1.c 	ms->timer_pending = 0;
ms               1129 drivers/md/dm-raid1.c 	INIT_WORK(&ms->trigger_event, trigger_event);
ms               1131 drivers/md/dm-raid1.c 	r = parse_features(ms, argc, argv, &args_used);
ms               1153 drivers/md/dm-raid1.c 	ms->kcopyd_client = dm_kcopyd_client_create(&dm_kcopyd_throttle);
ms               1154 drivers/md/dm-raid1.c 	if (IS_ERR(ms->kcopyd_client)) {
ms               1155 drivers/md/dm-raid1.c 		r = PTR_ERR(ms->kcopyd_client);
ms               1159 drivers/md/dm-raid1.c 	wakeup_mirrord(ms);
ms               1163 drivers/md/dm-raid1.c 	destroy_workqueue(ms->kmirrord_wq);
ms               1165 drivers/md/dm-raid1.c 	free_context(ms, ti, ms->nr_mirrors);
ms               1171 drivers/md/dm-raid1.c 	struct mirror_set *ms = (struct mirror_set *) ti->private;
ms               1173 drivers/md/dm-raid1.c 	del_timer_sync(&ms->timer);
ms               1174 drivers/md/dm-raid1.c 	flush_workqueue(ms->kmirrord_wq);
ms               1175 drivers/md/dm-raid1.c 	flush_work(&ms->trigger_event);
ms               1176 drivers/md/dm-raid1.c 	dm_kcopyd_client_destroy(ms->kcopyd_client);
ms               1177 drivers/md/dm-raid1.c 	destroy_workqueue(ms->kmirrord_wq);
ms               1178 drivers/md/dm-raid1.c 	free_context(ms, ti, ms->nr_mirrors);
ms               1188 drivers/md/dm-raid1.c 	struct mirror_set *ms = ti->private;
ms               1189 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms               1197 drivers/md/dm-raid1.c 		bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio);
ms               1198 drivers/md/dm-raid1.c 		queue_bio(ms, bio, rw);
ms               1202 drivers/md/dm-raid1.c 	r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0);
ms               1213 drivers/md/dm-raid1.c 		queue_bio(ms, bio, rw);
ms               1221 drivers/md/dm-raid1.c 	m = choose_mirror(ms, bio->bi_iter.bi_sector);
ms               1237 drivers/md/dm-raid1.c 	struct mirror_set *ms = (struct mirror_set *) ti->private;
ms               1249 drivers/md/dm-raid1.c 			dm_rh_dec(ms->rh, bio_record->write_region);
ms               1281 drivers/md/dm-raid1.c 		if (default_ok(m) || mirror_available(ms, bio)) {
ms               1288 drivers/md/dm-raid1.c 			queue_bio(ms, bio, rw);
ms               1302 drivers/md/dm-raid1.c 	struct mirror_set *ms = (struct mirror_set *) ti->private;
ms               1303 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms               1308 drivers/md/dm-raid1.c 	atomic_set(&ms->suspend, 1);
ms               1316 drivers/md/dm-raid1.c 	spin_lock_irq(&ms->lock);
ms               1317 drivers/md/dm-raid1.c 	holds = ms->holds;
ms               1318 drivers/md/dm-raid1.c 	bio_list_init(&ms->holds);
ms               1319 drivers/md/dm-raid1.c 	spin_unlock_irq(&ms->lock);
ms               1322 drivers/md/dm-raid1.c 		hold_bio(ms, bio);
ms               1328 drivers/md/dm-raid1.c 	dm_rh_stop_recovery(ms->rh);
ms               1331 drivers/md/dm-raid1.c 		   !dm_rh_recovery_in_flight(ms->rh));
ms               1343 drivers/md/dm-raid1.c 	flush_workqueue(ms->kmirrord_wq);
ms               1348 drivers/md/dm-raid1.c 	struct mirror_set *ms = ti->private;
ms               1349 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms               1358 drivers/md/dm-raid1.c 	struct mirror_set *ms = ti->private;
ms               1359 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms               1361 drivers/md/dm-raid1.c 	atomic_set(&ms->suspend, 0);
ms               1365 drivers/md/dm-raid1.c 	dm_rh_start_recovery(ms->rh);
ms               1398 drivers/md/dm-raid1.c 	struct mirror_set *ms = (struct mirror_set *) ti->private;
ms               1399 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
ms               1404 drivers/md/dm-raid1.c 		DMEMIT("%d ", ms->nr_mirrors);
ms               1405 drivers/md/dm-raid1.c 		for (m = 0; m < ms->nr_mirrors; m++) {
ms               1406 drivers/md/dm-raid1.c 			DMEMIT("%s ", ms->mirror[m].dev->name);
ms               1407 drivers/md/dm-raid1.c 			buffer[m] = device_status_char(&(ms->mirror[m]));
ms               1413 drivers/md/dm-raid1.c 		      (unsigned long long)ms->nr_regions, buffer);
ms               1422 drivers/md/dm-raid1.c 		DMEMIT("%d", ms->nr_mirrors);
ms               1423 drivers/md/dm-raid1.c 		for (m = 0; m < ms->nr_mirrors; m++)
ms               1424 drivers/md/dm-raid1.c 			DMEMIT(" %s %llu", ms->mirror[m].dev->name,
ms               1425 drivers/md/dm-raid1.c 			       (unsigned long long)ms->mirror[m].offset);
ms               1427 drivers/md/dm-raid1.c 		num_feature_args += !!errors_handled(ms);
ms               1428 drivers/md/dm-raid1.c 		num_feature_args += !!keep_log(ms);
ms               1431 drivers/md/dm-raid1.c 			if (errors_handled(ms))
ms               1433 drivers/md/dm-raid1.c 			if (keep_log(ms))
ms               1444 drivers/md/dm-raid1.c 	struct mirror_set *ms = ti->private;
ms               1448 drivers/md/dm-raid1.c 	for (i = 0; !ret && i < ms->nr_mirrors; i++)
ms               1449 drivers/md/dm-raid1.c 		ret = fn(ti, ms->mirror[i].dev,
ms               1450 drivers/md/dm-raid1.c 			 ms->mirror[i].offset, ti->len, data);
ms                173 drivers/media/dvb-frontends/horus3a.c 	u32 ms = 0;
ms                196 drivers/media/dvb-frontends/horus3a.c 	ms = DIV_ROUND_CLOSEST((frequency * mixdiv) / 2, 1000);
ms                197 drivers/media/dvb-frontends/horus3a.c 	if (ms > 0x7FFF) { /* 15 bit */
ms                292 drivers/media/dvb-frontends/horus3a.c 	data[0] = (u8)((ms >> 7) & 0xFF);
ms                293 drivers/media/dvb-frontends/horus3a.c 	data[1] = (u8)((ms << 1) & 0xFF);
ms                309 drivers/media/dvb-frontends/horus3a.c 	priv->frequency = ms * 2 * 1000 / mixdiv;
ms                300 drivers/media/i2c/m5mols/m5mols_core.c 	int ms = timeout < 0 ? M5MOLS_BUSY_WAIT_DEF_TIMEOUT : timeout;
ms                301 drivers/media/i2c/m5mols/m5mols_core.c 	unsigned long end = jiffies + msecs_to_jiffies(ms);
ms                312 drivers/media/i2c/m5mols/m5mols_core.c 	} while (ms > 0 && time_is_after_jiffies(end));
ms                278 drivers/media/pci/solo6x10/solo6x10-core.c 	unsigned long ms;
ms                279 drivers/media/pci/solo6x10/solo6x10-core.c 	int ret = kstrtoul(buf, 10, &ms);
ms                281 drivers/media/pci/solo6x10/solo6x10-core.c 	if (ret < 0 || ms > 200)
ms                283 drivers/media/pci/solo6x10/solo6x10-core.c 	solo_dev->p2m_jiffies = msecs_to_jiffies(ms);
ms                415 drivers/media/platform/vivid/vivid-kthread-cap.c 	unsigned ms;
ms                478 drivers/media/platform/vivid/vivid-kthread-cap.c 	ms = dev->ms_vid_cap;
ms                481 drivers/media/platform/vivid/vivid-kthread-cap.c 				(ms / (60 * 60 * 1000)) % 24,
ms                482 drivers/media/platform/vivid/vivid-kthread-cap.c 				(ms / (60 * 1000)) % 60,
ms                483 drivers/media/platform/vivid/vivid-kthread-cap.c 				(ms / 1000) % 60,
ms                484 drivers/media/platform/vivid/vivid-kthread-cap.c 				ms % 1000,
ms                 78 drivers/media/platform/vivid/vivid-radio-common.c 		rds->ms = dev->radio_tx_rds_ms->cur.val;
ms                 93 drivers/media/platform/vivid/vivid-radio-common.c 		v4l2_ctrl_s_ctrl(dev->radio_rx_rds_ms, rds->ms);
ms                 66 drivers/media/platform/vivid/vivid-rds-gen.c 			data[1].lsb |= (rds->ta << 4) | (rds->ms << 3);
ms                113 drivers/media/platform/vivid/vivid-rds-gen.c 			data[1].lsb |= (rds->ta << 4) | (rds->ms << 3);
ms                120 drivers/media/platform/vivid/vivid-rds-gen.c 			data[3].lsb |= (rds->ta << 4) | (rds->ms << 3);
ms                146 drivers/media/platform/vivid/vivid-rds-gen.c 	rds->ms = true;
ms                 33 drivers/media/platform/vivid/vivid-rds-gen.h 	bool			ms;
ms                977 drivers/mmc/core/block.c 	unsigned int ms = DIV_ROUND_UP(data->timeout_ns, 1000000);
ms                982 drivers/mmc/core/block.c 		ms += DIV_ROUND_UP(data->timeout_clks, khz);
ms                985 drivers/mmc/core/block.c 	return ms;
ms                 61 drivers/mmc/core/core.h static inline void mmc_delay(unsigned int ms)
ms                 63 drivers/mmc/core/core.h 	if (ms <= 20)
ms                 64 drivers/mmc/core/core.h 		usleep_range(ms * 1000, ms * 1250);
ms                 66 drivers/mmc/core/core.h 		msleep(ms);
ms                336 drivers/mmc/host/mxs-mmc.c 	const unsigned int ms = ns / 1000;
ms                337 drivers/mmc/host/mxs-mmc.c 	const unsigned int ticks = ms * clock_per_ms;
ms                164 drivers/mtd/tests/speedtest.c 	long ms;
ms                166 drivers/mtd/tests/speedtest.c 	ms = ktime_ms_delta(finish, start);
ms                167 drivers/mtd/tests/speedtest.c 	if (ms == 0)
ms                170 drivers/mtd/tests/speedtest.c 	do_div(k, ms);
ms                322 drivers/mtd/tests/torturetest.c 			long ms;
ms                325 drivers/mtd/tests/torturetest.c 			ms = ktime_ms_delta(finish, start);
ms                328 drivers/mtd/tests/torturetest.c 			       erase_cycles, ms, ms / 1000);
ms                 21 drivers/net/dsa/sja1105/sja1105.h #define SJA1105_AGEING_TIME_MS(ms)	((ms) / 10)
ms               2063 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h static inline u32 reg_poll(struct bnx2x *bp, u32 reg, u32 expected, int ms,
ms               2072 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h 		ms -= wait;
ms               2075 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h 	} while (ms > 0);
ms               1400 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	u32 ms;
ms               1405 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	for (ms = 0; (ret != 0) && ((*timeout == 0) || (ms <= *timeout));
ms               1406 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	     ms += HZ / 10) {
ms                294 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	int i, ms, delay_idx, ret;
ms                327 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	ms = delay[0];
ms                329 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	for (i = 0; ; i += ms) {
ms                354 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			ms = delay[delay_idx];  /* last element may repeat */
ms                357 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			msleep(ms);
ms                359 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			mdelay(ms);
ms                387 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	ms = delay[0];
ms                392 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	     i += ms) {
ms                394 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			ms = delay[delay_idx];  /* last element may repeat */
ms                397 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			msleep(ms);
ms                399 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			mdelay(ms);
ms                420 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			execute = i + ms;
ms               7131 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		int ms;
ms               7134 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		for (ms = 0; ms < FW_CMD_MAX_TIMEOUT; ) {
ms               7138 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			ms += 100;
ms                137 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	int i, ms, delay_idx, ret;
ms                170 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	ms = delay[0];
ms                172 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	for (i = 0; ; i += ms) {
ms                196 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			ms = delay[delay_idx];  /* last element may repeat */
ms                199 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			msleep(ms);
ms                201 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			mdelay(ms);
ms                248 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	ms = delay[0];
ms                250 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	for (i = 0; i < FW_CMD_MAX_TIMEOUT; i += ms) {
ms                252 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			ms = delay[delay_idx];
ms                255 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			msleep(ms);
ms                257 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			mdelay(ms);
ms                296 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			execute = i + ms;
ms                268 drivers/net/ethernet/davicom/dm9000.c static void dm9000_msleep(struct board_info *db, unsigned int ms)
ms                271 drivers/net/ethernet/davicom/dm9000.c 		mdelay(ms);
ms                273 drivers/net/ethernet/davicom/dm9000.c 		msleep(ms);
ms               1303 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 				   struct qlcnic_ms_reg_ctrl *ms)
ms               1305 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	ms->control = QLCNIC_MS_CTRL;
ms               1306 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	ms->low = QLCNIC_MS_ADDR_LO;
ms               1307 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	ms->hi = QLCNIC_MS_ADDR_HI;
ms               1309 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[0] = QLCNIC_MS_WRTDATA_LO;
ms               1310 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[0] = QLCNIC_MS_RDDATA_LO;
ms               1311 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[1] = QLCNIC_MS_WRTDATA_HI;
ms               1312 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[1] = QLCNIC_MS_RDDATA_HI;
ms               1313 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[2] = QLCNIC_MS_WRTDATA_ULO;
ms               1314 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[3] = QLCNIC_MS_WRTDATA_UHI;
ms               1315 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[2] = QLCNIC_MS_RDDATA_ULO;
ms               1316 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[3] = QLCNIC_MS_RDDATA_UHI;
ms               1318 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[0] = QLCNIC_MS_WRTDATA_ULO;
ms               1319 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[0] = QLCNIC_MS_RDDATA_ULO;
ms               1320 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[1] = QLCNIC_MS_WRTDATA_UHI;
ms               1321 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[1] = QLCNIC_MS_RDDATA_UHI;
ms               1322 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[2] = QLCNIC_MS_WRTDATA_LO;
ms               1323 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->wd[3] = QLCNIC_MS_WRTDATA_HI;
ms               1324 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[2] = QLCNIC_MS_RDDATA_LO;
ms               1325 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		ms->rd[3] = QLCNIC_MS_RDDATA_HI;
ms               1328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	ms->ocm_window = OCM_WIN_P3P(off);
ms               1329 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	ms->off = GET_MEM_OFFS_2M(off);
ms               1336 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_ms_reg_ctrl ms;
ms               1342 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl));
ms               1349 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_set_ms_controls(adapter, off, &ms);
ms               1352 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window,
ms               1353 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 						    ms.off, &data, 1);
ms               1359 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.low, off8);
ms               1360 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.hi, 0);
ms               1362 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE);
ms               1363 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE);
ms               1366 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		temp = qlcnic_ind_rd(adapter, ms.control);
ms               1377 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.wd[0], qlcnic_ind_rd(adapter, ms.rd[0]));
ms               1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.wd[1], qlcnic_ind_rd(adapter, ms.rd[1]));
ms               1380 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.wd[2], data & 0xffffffff);
ms               1381 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.wd[3], (data >> 32) & 0xffffffff);
ms               1383 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_ENABLE);
ms               1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_START);
ms               1387 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		temp = qlcnic_ind_rd(adapter, ms.control);
ms               1411 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_ms_reg_ctrl ms;
ms               1422 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl));
ms               1423 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_set_ms_controls(adapter, off, &ms);
ms               1426 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window,
ms               1427 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 						    ms.off, data, 0);
ms               1433 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.low, off8);
ms               1434 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.hi, 0);
ms               1436 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE);
ms               1437 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE);
ms               1440 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		temp = qlcnic_ind_rd(adapter, ms.control);
ms               1452 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		temp = qlcnic_ind_rd(adapter, ms.rd[3]);
ms               1454 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		val |= qlcnic_ind_rd(adapter, ms.rd[2]);
ms                 45 drivers/net/fddi/defxx.h 	PI_UINT32  ms;
ms                280 drivers/net/hamradio/yam.c static void delay(int ms)
ms                282 drivers/net/hamradio/yam.c 	unsigned long timeout = jiffies + ((ms * HZ) / 1000);
ms                669 drivers/net/ieee802154/ca8210.c static void ca8210_reset_send(struct spi_device *spi, unsigned int ms)
ms                677 drivers/net/ieee802154/ca8210.c 	msleep(ms);
ms                389 drivers/net/wimax/i2400m/control.c 				    const struct i2400m_tlv_media_status *ms)
ms                394 drivers/net/wimax/i2400m/control.c 	enum i2400m_media_status status = le32_to_cpu(ms->media_status);
ms                396 drivers/net/wimax/i2400m/control.c 	d_fnstart(3, dev, "(i2400m %p ms %p [%u])\n", i2400m, ms, status);
ms                418 drivers/net/wimax/i2400m/control.c 		i2400m, ms, status);
ms                438 drivers/net/wimax/i2400m/control.c 	const struct i2400m_tlv_media_status *ms;
ms                459 drivers/net/wimax/i2400m/control.c 	if (0 == i2400m_tlv_match(tlv, I2400M_TLV_MEDIA_STATUS, sizeof(*ms))) {
ms                460 drivers/net/wimax/i2400m/control.c 		ms = container_of(tlv, typeof(*ms), hdr);
ms                462 drivers/net/wimax/i2400m/control.c 			 tag, le32_to_cpu(ms->media_status));
ms                463 drivers/net/wimax/i2400m/control.c 		i2400m_report_tlv_media_status(i2400m, ms);
ms                955 drivers/net/wimax/i2400m/i2400m.h void __i2400m_msleep(unsigned ms)
ms                959 drivers/net/wimax/i2400m/i2400m.h 	msleep(ms);
ms                235 drivers/net/wireless/ath/ath9k/channel.c 	u64 ms;
ms                240 drivers/net/wireless/ath/ath9k/channel.c 	ms = ts.tv_sec * 1000 + ts.tv_nsec / 1000000;
ms                241 drivers/net/wireless/ath/ath9k/channel.c 	ms -= old->tv_sec * 1000 + old->tv_nsec / 1000000;
ms                244 drivers/net/wireless/ath/ath9k/channel.c 	return (u32)ms;
ms                620 drivers/net/wireless/ath/carl9170/rx.c static bool carl9170_ampdu_check(struct ar9170 *ar, u8 *buf, u8 ms,
ms                625 drivers/net/wireless/ath/carl9170/rx.c 	if ((ms & AR9170_RX_STATUS_MPDU) == AR9170_RX_STATUS_MPDU_SINGLE) {
ms               2502 drivers/net/wireless/ath/wil6210/cfg80211.c 		struct cfg80211_match_set *ms = &request->match_sets[i];
ms               2505 drivers/net/wireless/ath/wil6210/cfg80211.c 			     i, ms->rssi_thold);
ms               2507 drivers/net/wireless/ath/wil6210/cfg80211.c 				  ms->ssid.ssid,
ms               2508 drivers/net/wireless/ath/wil6210/cfg80211.c 				  ms->ssid.ssid_len, true);
ms               3465 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c static __always_inline void brcmf_delay(u32 ms)
ms               3467 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 	if (ms < 1000 / HZ) {
ms               3469 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		mdelay(ms);
ms               3471 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		msleep(ms);
ms                363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 	struct cfg80211_match_set *ms;
ms                371 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			ms = &r->match_sets[j];
ms                372 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			if (ms->ssid.ssid_len) {
ms                373 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 				active = brcmf_is_ssid_active(&ms->ssid, r);
ms                374 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 				err = brcmf_pno_add_ssid(ifp, &ms->ssid,
ms                377 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			if (!err && is_valid_ether_addr(ms->bssid))
ms                378 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 				err = brcmf_pno_add_bssid(ifp, ms->bssid);
ms                565 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 	struct cfg80211_match_set *ms;
ms                576 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			ms = &req->match_sets[j];
ms                577 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			if (ms->ssid.ssid_len == ni->SSID_len &&
ms                578 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			    !memcmp(ms->ssid.ssid, ni->SSID, ni->SSID_len)) {
ms                582 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			if (is_valid_ether_addr(ms->bssid) &&
ms                583 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pno.c 			    !memcmp(ms->bssid, ni->bssid, ETH_ALEN)) {
ms                 43 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h 		__field(uint, ms)
ms                 48 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h 		__entry->ms = t->ms;
ms                 54 drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h 		__entry->ms, __entry->set, __entry->periodic
ms               1468 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 						     msecs_to_jiffies(t->ms));
ms               1517 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c void brcms_add_timer(struct brcms_timer *t, uint ms, int periodic)
ms               1527 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	t->ms = ms;
ms               1534 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	ieee80211_queue_delayed_work(hw, &t->dly_wrk, msecs_to_jiffies(ms));
ms                 41 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h 	uint ms;
ms                107 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h void brcms_add_timer(struct brcms_timer *timer, uint ms, int periodic);
ms                 74 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic)
ms                 76 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c 	brcms_add_timer((struct brcms_timer *)t, ms, periodic);
ms                137 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h void wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic);
ms               1364 drivers/net/wireless/intel/iwlegacy/common.c il_scan_cancel_timeout(struct il_priv *il, unsigned long ms)
ms               1366 drivers/net/wireless/intel/iwlegacy/common.c 	unsigned long timeout = jiffies + msecs_to_jiffies(ms);
ms               1773 drivers/net/wireless/intel/iwlegacy/common.h int il_scan_cancel_timeout(struct il_priv *il, unsigned long ms);
ms                247 drivers/net/wireless/intel/iwlwifi/dvm/agn.h void iwl_scan_cancel_timeout(struct iwl_priv *priv, unsigned long ms);
ms                204 drivers/net/wireless/intel/iwlwifi/dvm/scan.c void iwl_scan_cancel_timeout(struct iwl_priv *priv, unsigned long ms)
ms                206 drivers/net/wireless/intel/iwlwifi/dvm/scan.c 	unsigned long timeout = jiffies + msecs_to_jiffies(ms);
ms               2269 drivers/net/wireless/marvell/mwl8k.c 		int ms;
ms               2271 drivers/net/wireless/marvell/mwl8k.c 		ms = MWL8K_CMD_TIMEOUT_MS - jiffies_to_msecs(timeout);
ms               2278 drivers/net/wireless/marvell/mwl8k.c 		else if (ms > 2000)
ms               2282 drivers/net/wireless/marvell/mwl8k.c 				     ms);
ms                215 drivers/nvme/host/core.c 	return ns->pi_type && ns->ms == sizeof(struct t10_pi_tuple);
ms                707 drivers/nvme/host/core.c 	if (ns->ms) {
ms               1291 drivers/nvme/host/core.c 	meta_len = (io.nblocks + 1) * ns->ms;
ms               1645 drivers/nvme/host/core.c static void nvme_init_integrity(struct gendisk *disk, u16 ms, u8 pi_type)
ms               1666 drivers/nvme/host/core.c 	integrity.tuple_size = ms;
ms               1671 drivers/nvme/host/core.c static void nvme_init_integrity(struct gendisk *disk, u16 ms, u8 pi_type)
ms               1813 drivers/nvme/host/core.c 	if (ns->ms && !ns->ext &&
ms               1815 drivers/nvme/host/core.c 		nvme_init_integrity(disk, ns->ms, ns->pi_type);
ms               1816 drivers/nvme/host/core.c 	if ((ns->ms && !nvme_ns_has_pi(ns) && !blk_get_integrity(disk)) ||
ms               1845 drivers/nvme/host/core.c 	ns->ms = le16_to_cpu(id->lbaf[id->flbas & NVME_NS_FLBAS_LBA_MASK].ms);
ms               1846 drivers/nvme/host/core.c 	ns->ext = ns->ms && (id->flbas & NVME_NS_FLBAS_META_EXT);
ms               1848 drivers/nvme/host/core.c 	if (ns->ms == sizeof(struct t10_pi_tuple))
ms                963 drivers/nvme/host/lightnvm.c 	geo->sos = ns->ms;
ms                368 drivers/nvme/host/nvme.h 	u16 ms;
ms                880 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c DB8500_FUNC_GROUPS(ms, "ms_c_1");
ms                939 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c 	FUNCTION(ms),
ms                 55 drivers/pwm/pwm-lpss.c 	const unsigned int ms = 500 * USEC_PER_MSEC;
ms                 70 drivers/pwm/pwm-lpss.c 	err = readl_poll_timeout(addr, val, !(val & PWM_SW_UPDATE), 40, ms);
ms               2319 drivers/regulator/core.c 	unsigned int ms = delay / 1000;
ms               2322 drivers/regulator/core.c 	if (ms > 0) {
ms               2327 drivers/regulator/core.c 		if (ms < 20)
ms               2328 drivers/regulator/core.c 			us += ms * 1000;
ms               2330 drivers/regulator/core.c 			msleep(ms);
ms               2784 drivers/regulator/core.c int regulator_disable_deferred(struct regulator *regulator, int ms)
ms               2788 drivers/regulator/core.c 	if (!ms)
ms               2794 drivers/regulator/core.c 			 msecs_to_jiffies(ms));
ms                421 drivers/remoteproc/qcom_q6v5_mss.c static int q6v5_rmb_pbl_wait(struct q6v5 *qproc, int ms)
ms                426 drivers/remoteproc/qcom_q6v5_mss.c 	timeout = jiffies + msecs_to_jiffies(ms);
ms                441 drivers/remoteproc/qcom_q6v5_mss.c static int q6v5_rmb_mba_wait(struct q6v5 *qproc, u32 status, int ms)
ms                447 drivers/remoteproc/qcom_q6v5_mss.c 	timeout = jiffies + msecs_to_jiffies(ms);
ms                225 drivers/s390/scsi/zfcp_fc.c 	zfcp_fc_wka_port_force_offline(&gs->ms);
ms                977 drivers/s390/scsi/zfcp_fc.c 		return &adapter->gs->ms;
ms               1091 drivers/s390/scsi/zfcp_fc.c 	zfcp_fc_wka_port_init(&wka_ports->ms, FC_FID_MGMT_SERV, adapter);
ms                215 drivers/s390/scsi/zfcp_fc.h 	struct zfcp_fc_wka_port ms;
ms                665 drivers/scsi/aic94xx/aic94xx_sds.c static int asd_validate_ms(struct asd_manuf_sec *ms)
ms                667 drivers/scsi/aic94xx/aic94xx_sds.c 	if (ms->sig[0] != 'S' || ms->sig[1] != 'M') {
ms                669 drivers/scsi/aic94xx/aic94xx_sds.c 			    ms->sig[0], ms->sig[1]);
ms                672 drivers/scsi/aic94xx/aic94xx_sds.c 	if (ms->maj != 0) {
ms                674 drivers/scsi/aic94xx/aic94xx_sds.c 			   ms->maj);
ms                677 drivers/scsi/aic94xx/aic94xx_sds.c 	ms->offs_next = le16_to_cpu((__force __le16) ms->offs_next);
ms                678 drivers/scsi/aic94xx/aic94xx_sds.c 	ms->chksum = le16_to_cpu((__force __le16) ms->chksum);
ms                679 drivers/scsi/aic94xx/aic94xx_sds.c 	ms->size = le16_to_cpu((__force __le16) ms->size);
ms                681 drivers/scsi/aic94xx/aic94xx_sds.c 	if (asd_calc_flash_chksum((u16 *)ms, ms->size/2)) {
ms                689 drivers/scsi/aic94xx/aic94xx_sds.c 			       struct asd_manuf_sec *ms)
ms                691 drivers/scsi/aic94xx/aic94xx_sds.c 	memcpy(asd_ha->hw_prof.sas_addr, ms->sas_addr, SAS_ADDR_SIZE);
ms                696 drivers/scsi/aic94xx/aic94xx_sds.c 			      struct asd_manuf_sec *ms)
ms                698 drivers/scsi/aic94xx/aic94xx_sds.c 	memcpy(asd_ha->hw_prof.pcba_sn, ms->pcba_sn, ASD_PCBA_SN_SIZE);
ms                 85 drivers/scsi/bfa/bfa_fcs.h 	struct bfa_fcs_lport_ms_s *ms;	/*  parent ms */
ms                108 drivers/scsi/bfa/bfa_fcs.h 	struct bfa_fcs_lport_ms_s ms;	/*  MS component of port */
ms                157 drivers/scsi/bfa/bfa_fcs.h #define BFA_FCS_GET_MS_FROM_PORT(port)  (&port->port_topo.pfab.ms)
ms                158 drivers/scsi/bfa/bfa_fcs.h #define BFA_FCS_GET_FDMI_FROM_PORT(port)  (&port->port_topo.pfab.ms.fdmi)
ms                310 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_init(struct bfa_fcs_lport_ms_s *ms);
ms                311 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_offline(struct bfa_fcs_lport_ms_s *ms);
ms                312 drivers/scsi/bfa/bfa_fcs.h void bfa_fcs_lport_fdmi_online(struct bfa_fcs_lport_ms_s *ms);
ms               1481 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1521 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1546 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1595 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1626 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1651 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1698 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1729 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1754 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1799 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1827 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1848 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1863 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               1902 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2162 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2196 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2462 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2485 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2519 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2556 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2577 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2616 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2672 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = fdmi->ms->port;
ms               2791 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_init(struct bfa_fcs_lport_ms_s *ms)
ms               2793 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi;
ms               2795 drivers/scsi/bfa/bfa_fcs_lport.c 	fdmi->ms = ms;
ms               2796 drivers/scsi/bfa/bfa_fcs_lport.c 	if (ms->port->fcs->fdmi_enabled)
ms               2803 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_offline(struct bfa_fcs_lport_ms_s *ms)
ms               2805 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi;
ms               2807 drivers/scsi/bfa/bfa_fcs_lport.c 	fdmi->ms = ms;
ms               2812 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_fdmi_online(struct bfa_fcs_lport_ms_s *ms)
ms               2814 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi;
ms               2816 drivers/scsi/bfa/bfa_fcs_lport.c 	fdmi->ms = ms;
ms               2871 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms,
ms               2873 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms,
ms               2875 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms,
ms               2877 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms,
ms               2879 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms,
ms               2881 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms,
ms               2883 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms,
ms               2885 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms,
ms               2887 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms,
ms               2889 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms,
ms               2891 drivers/scsi/bfa/bfa_fcs_lport.c static void     bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms,
ms               2897 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms,
ms               2900 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               2901 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               2905 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending);
ms               2906 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_plogi(ms, NULL);
ms               2913 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               2918 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms,
ms               2921 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               2922 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               2926 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi);
ms               2930 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               2931 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               2932 drivers/scsi/bfa/bfa_fcs_lport.c 					   &ms->fcxp_wqe);
ms               2936 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               2941 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms,
ms               2944 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               2945 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               2952 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_retry);
ms               2953 drivers/scsi/bfa/bfa_fcs_lport.c 		ms->port->stats.ms_retries++;
ms               2954 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               2955 drivers/scsi/bfa/bfa_fcs_lport.c 				    &ms->timer, bfa_fcs_lport_ms_timeout, ms,
ms               2963 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_fdmi_online(ms);
ms               2968 drivers/scsi/bfa/bfa_fcs_lport.c 		if (ms->port->vport) {
ms               2969 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online);
ms               2977 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending);
ms               2978 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_gmal(ms, NULL);
ms               2982 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               2983 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_discard(ms->fcxp);
ms               2987 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               2992 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms,
ms               2995 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               2996 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3003 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending);
ms               3004 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_plogi(ms, NULL);
ms               3008 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3009 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_timer_stop(&ms->timer);
ms               3013 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3018 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms,
ms               3021 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3022 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3026 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3030 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending);
ms               3031 drivers/scsi/bfa/bfa_fcs_lport.c 		ms->retry_cnt = 0;
ms               3032 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_gfn(ms, NULL);
ms               3036 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3041 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms,
ms               3044 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3045 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3049 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal);
ms               3053 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3054 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               3055 drivers/scsi/bfa/bfa_fcs_lport.c 					   &ms->fcxp_wqe);
ms               3059 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3064 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms,
ms               3067 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3068 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3075 drivers/scsi/bfa/bfa_fcs_lport.c 		if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) {
ms               3076 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_retry);
ms               3077 drivers/scsi/bfa/bfa_fcs_lport.c 			ms->port->stats.ms_retries++;
ms               3078 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               3079 drivers/scsi/bfa/bfa_fcs_lport.c 				&ms->timer, bfa_fcs_lport_ms_timeout, ms,
ms               3082 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending);
ms               3083 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_fcs_lport_ms_send_gfn(ms, NULL);
ms               3084 drivers/scsi/bfa/bfa_fcs_lport.c 			ms->retry_cnt = 0;
ms               3089 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending);
ms               3090 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_gfn(ms, NULL);
ms               3094 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3095 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_discard(ms->fcxp);
ms               3099 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3104 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms,
ms               3107 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3108 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3115 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending);
ms               3116 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_gmal(ms, NULL);
ms               3120 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3121 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_timer_stop(&ms->timer);
ms               3125 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3135 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = ms_cbarg;
ms               3136 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_t *port = ms->port;
ms               3146 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe,
ms               3147 drivers/scsi/bfa/bfa_fcs_lport.c 				bfa_fcs_lport_ms_send_gmal, ms, BFA_TRUE);
ms               3150 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->fcxp = fcxp;
ms               3158 drivers/scsi/bfa/bfa_fcs_lport.c 			  bfa_fcs_lport_ms_gmal_response, (void *)ms,
ms               3161 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT);
ms               3170 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg;
ms               3171 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_t *port = ms->port;
ms               3186 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3198 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3234 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK);
ms               3240 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3244 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms,
ms               3247 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3248 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3252 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn);
ms               3256 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3257 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               3258 drivers/scsi/bfa/bfa_fcs_lport.c 					   &ms->fcxp_wqe);
ms               3262 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3267 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms,
ms               3270 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3271 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3278 drivers/scsi/bfa/bfa_fcs_lport.c 		if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) {
ms               3279 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_retry);
ms               3280 drivers/scsi/bfa/bfa_fcs_lport.c 			ms->port->stats.ms_retries++;
ms               3281 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port),
ms               3282 drivers/scsi/bfa/bfa_fcs_lport.c 				&ms->timer, bfa_fcs_lport_ms_timeout, ms,
ms               3285 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online);
ms               3286 drivers/scsi/bfa/bfa_fcs_lport.c 			ms->retry_cnt = 0;
ms               3291 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online);
ms               3295 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3296 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcxp_discard(ms->fcxp);
ms               3300 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3305 drivers/scsi/bfa/bfa_fcs_lport.c bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms,
ms               3308 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn);
ms               3309 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_trc(ms->port->fcs, event);
ms               3316 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending);
ms               3317 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_lport_ms_send_gfn(ms, NULL);
ms               3321 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3322 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_timer_stop(&ms->timer);
ms               3326 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_fault(ms->port->fcs, event);
ms               3336 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = ms_cbarg;
ms               3337 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_t *port = ms->port;
ms               3347 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe,
ms               3348 drivers/scsi/bfa/bfa_fcs_lport.c 				bfa_fcs_lport_ms_send_gfn, ms, BFA_TRUE);
ms               3351 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->fcxp = fcxp;
ms               3359 drivers/scsi/bfa/bfa_fcs_lport.c 			  bfa_fcs_lport_ms_gfn_response, (void *)ms,
ms               3362 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT);
ms               3370 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg;
ms               3371 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_t *port = ms->port;
ms               3383 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3397 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK);
ms               3403 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3413 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = ms_cbarg;
ms               3414 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = ms->port;
ms               3425 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe,
ms               3426 drivers/scsi/bfa/bfa_fcs_lport.c 				bfa_fcs_lport_ms_send_plogi, ms, BFA_TRUE);
ms               3429 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->fcxp = fcxp;
ms               3440 drivers/scsi/bfa/bfa_fcs_lport.c 			  bfa_fcs_lport_ms_plogi_response, (void *)ms,
ms               3444 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT);
ms               3452 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg;
ms               3453 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_s *port = ms->port;
ms               3466 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3478 drivers/scsi/bfa/bfa_fcs_lport.c 			bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3482 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK);
ms               3492 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3498 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR);
ms               3505 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) arg;
ms               3507 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->port->stats.ms_timeouts++;
ms               3508 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_TIMEOUT);
ms               3515 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port);
ms               3517 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->port = port;
ms               3518 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline);
ms               3523 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_fdmi_init(ms);
ms               3529 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port);
ms               3531 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->port = port;
ms               3532 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_PORT_OFFLINE);
ms               3533 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_fcs_lport_fdmi_offline(ms);
ms               3539 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port);
ms               3541 drivers/scsi/bfa/bfa_fcs_lport.c 	ms->port = port;
ms               3542 drivers/scsi/bfa/bfa_fcs_lport.c 	bfa_sm_send_event(ms, MSSM_EVENT_PORT_ONLINE);
ms               3547 drivers/scsi/bfa/bfa_fcs_lport.c 	struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port);
ms               3550 drivers/scsi/bfa/bfa_fcs_lport.c 	if (bfa_sm_cmp_state(ms, bfa_fcs_lport_ms_sm_online))
ms               3551 drivers/scsi/bfa/bfa_fcs_lport.c 		bfa_sm_send_event(ms, MSSM_EVENT_PORT_FABRIC_RSCN);
ms               1286 drivers/scsi/csiostor/csio_hw.c 		int ms;
ms               1289 drivers/scsi/csiostor/csio_hw.c 		for (ms = 0; ms < FW_CMD_MAX_TIMEOUT; ) {
ms               1293 drivers/scsi/csiostor/csio_hw.c 			ms += 100;
ms                151 drivers/scsi/lpfc/lpfc_debugfs.c 	uint32_t ms;
ms                169 drivers/scsi/lpfc/lpfc_debugfs.c 		ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time);
ms                172 drivers/scsi/lpfc/lpfc_debugfs.c 			dtp->seq_cnt, ms, dtp->fmt);
ms                180 drivers/scsi/lpfc/lpfc_debugfs.c 		ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time);
ms                183 drivers/scsi/lpfc/lpfc_debugfs.c 			dtp->seq_cnt, ms, dtp->fmt);
ms                217 drivers/scsi/lpfc/lpfc_debugfs.c 	uint32_t ms;
ms                235 drivers/scsi/lpfc/lpfc_debugfs.c 		ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time);
ms                238 drivers/scsi/lpfc/lpfc_debugfs.c 			dtp->seq_cnt, ms, dtp->fmt);
ms                246 drivers/scsi/lpfc/lpfc_debugfs.c 		ms = jiffies_to_msecs(dtp->jif - lpfc_debugfs_start_time);
ms                249 drivers/scsi/lpfc/lpfc_debugfs.c 			dtp->seq_cnt, ms, dtp->fmt);
ms                186 drivers/scsi/mesh.c static void mesh_done(struct mesh_state *ms, int start_next);
ms                187 drivers/scsi/mesh.c static void mesh_interrupt(struct mesh_state *ms);
ms                188 drivers/scsi/mesh.c static void cmd_complete(struct mesh_state *ms);
ms                189 drivers/scsi/mesh.c static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd);
ms                190 drivers/scsi/mesh.c static void halt_dma(struct mesh_state *ms);
ms                191 drivers/scsi/mesh.c static void phase_mismatch(struct mesh_state *ms);
ms                213 drivers/scsi/mesh.c static void dlog(struct mesh_state *ms, char *fmt, int a)
ms                215 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms                219 drivers/scsi/mesh.c 	slp = &ms->log[ms->log_ix];
ms                222 drivers/scsi/mesh.c 	tlp->phase = (ms->msgphase << 4) + ms->phase;
ms                223 drivers/scsi/mesh.c 	tlp->bs0 = ms->mesh->bus_status0;
ms                224 drivers/scsi/mesh.c 	tlp->bs1 = ms->mesh->bus_status1;
ms                225 drivers/scsi/mesh.c 	tlp->tgt = ms->conn_tgt;
ms                232 drivers/scsi/mesh.c 	if (++ms->log_ix >= N_DBG_SLOG)
ms                233 drivers/scsi/mesh.c 		ms->log_ix = 0;
ms                234 drivers/scsi/mesh.c 	if (ms->n_log < N_DBG_SLOG)
ms                235 drivers/scsi/mesh.c 		++ms->n_log;
ms                238 drivers/scsi/mesh.c static void dumplog(struct mesh_state *ms, int t)
ms                240 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[t];
ms                264 drivers/scsi/mesh.c static void dumpslog(struct mesh_state *ms)
ms                269 drivers/scsi/mesh.c 	if (ms->n_log == 0)
ms                271 drivers/scsi/mesh.c 	i = ms->log_ix - ms->n_log;
ms                274 drivers/scsi/mesh.c 	ms->n_log = 0;
ms                276 drivers/scsi/mesh.c 		lp = &ms->log[i];
ms                286 drivers/scsi/mesh.c 	} while (i != ms->log_ix);
ms                291 drivers/scsi/mesh.c static inline void dlog(struct mesh_state *ms, char *fmt, int a)
ms                293 drivers/scsi/mesh.c static inline void dumplog(struct mesh_state *ms, int tgt)
ms                295 drivers/scsi/mesh.c static inline void dumpslog(struct mesh_state *ms)
ms                303 drivers/scsi/mesh.c mesh_dump_regs(struct mesh_state *ms)
ms                305 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                306 drivers/scsi/mesh.c 	volatile struct dbdma_regs __iomem *md = ms->dma;
ms                311 drivers/scsi/mesh.c 	       ms, mr, md);
ms                323 drivers/scsi/mesh.c 	       ms->phase, ms->msgphase, ms->conn_tgt, ms->data_ptr);
ms                325 drivers/scsi/mesh.c 	       ms->dma_started, ms->dma_count, ms->n_msgout);
ms                327 drivers/scsi/mesh.c 		tp = &ms->tgts[t];
ms                348 drivers/scsi/mesh.c static void mesh_completed(struct mesh_state *ms, struct scsi_cmnd *cmd)
ms                358 drivers/scsi/mesh.c static void mesh_init(struct mesh_state *ms)
ms                360 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                361 drivers/scsi/mesh.c 	volatile struct dbdma_regs __iomem *md = ms->dma;
ms                374 drivers/scsi/mesh.c 	out_8(&mr->source_id, ms->host->this_id);
ms                400 drivers/scsi/mesh.c 	ms->phase = idle;
ms                401 drivers/scsi/mesh.c 	ms->msgphase = msg_none;
ms                405 drivers/scsi/mesh.c static void mesh_start_cmd(struct mesh_state *ms, struct scsi_cmnd *cmd)
ms                407 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                411 drivers/scsi/mesh.c 	ms->current_req = cmd;
ms                412 drivers/scsi/mesh.c 	ms->tgts[id].data_goes_out = cmd->sc_data_direction == DMA_TO_DEVICE;
ms                413 drivers/scsi/mesh.c 	ms->tgts[id].current_req = cmd;
ms                425 drivers/scsi/mesh.c 	if (ms->dma_started)
ms                428 drivers/scsi/mesh.c 	ms->phase = arbitrating;
ms                429 drivers/scsi/mesh.c 	ms->msgphase = msg_none;
ms                430 drivers/scsi/mesh.c 	ms->data_ptr = 0;
ms                431 drivers/scsi/mesh.c 	ms->dma_started = 0;
ms                432 drivers/scsi/mesh.c 	ms->n_msgout = 0;
ms                433 drivers/scsi/mesh.c 	ms->last_n_msgout = 0;
ms                434 drivers/scsi/mesh.c 	ms->expect_reply = 0;
ms                435 drivers/scsi/mesh.c 	ms->conn_tgt = id;
ms                436 drivers/scsi/mesh.c 	ms->tgts[id].saved_ptr = 0;
ms                437 drivers/scsi/mesh.c 	ms->stat = DID_OK;
ms                438 drivers/scsi/mesh.c 	ms->aborting = 0;
ms                440 drivers/scsi/mesh.c 	ms->tgts[id].n_log = 0;
ms                441 drivers/scsi/mesh.c 	dlog(ms, "start cmd=%x", (int) cmd);
ms                445 drivers/scsi/mesh.c 	dlog(ms, "about to arb, intr/exc/err/fc=%.8x",
ms                457 drivers/scsi/mesh.c 		dlog(ms, "busy b4 arb, intr/exc/err/fc=%.8x",
ms                464 drivers/scsi/mesh.c 				dlog(ms, "intr b4 arb, intr/exc/err/fc=%.8x",
ms                467 drivers/scsi/mesh.c 				mesh_interrupt(ms);
ms                468 drivers/scsi/mesh.c 				if (ms->phase != arbitrating)
ms                475 drivers/scsi/mesh.c 			ms->stat = DID_BUS_BUSY;
ms                476 drivers/scsi/mesh.c 			ms->phase = idle;
ms                477 drivers/scsi/mesh.c 			mesh_done(ms, 0);
ms                502 drivers/scsi/mesh.c 		dlog(ms, "intr after disresel, intr/exc/err/fc=%.8x",
ms                505 drivers/scsi/mesh.c 		mesh_interrupt(ms);
ms                506 drivers/scsi/mesh.c 		if (ms->phase != arbitrating)
ms                508 drivers/scsi/mesh.c 		dlog(ms, "after intr after disresel, intr/exc/err/fc=%.8x",
ms                520 drivers/scsi/mesh.c 	dlog(ms, "after arb, intr/exc/err/fc=%.8x",
ms                525 drivers/scsi/mesh.c 		dlog(ms, "resel? after arb, intr/exc/err/fc=%.8x",
ms                536 drivers/scsi/mesh.c 		dlog(ms, "tried reset after arb, intr/exc/err/fc=%.8x",
ms                558 drivers/scsi/mesh.c static void mesh_start(struct mesh_state *ms)
ms                562 drivers/scsi/mesh.c 	if (ms->phase != idle || ms->current_req != NULL) {
ms                564 drivers/scsi/mesh.c 		       ms->phase, ms);
ms                568 drivers/scsi/mesh.c 	while (ms->phase == idle) {
ms                570 drivers/scsi/mesh.c 		for (cmd = ms->request_q; ; cmd = (struct scsi_cmnd *) cmd->host_scribble) {
ms                573 drivers/scsi/mesh.c 			if (ms->tgts[cmd->device->id].current_req == NULL)
ms                579 drivers/scsi/mesh.c 			ms->request_q = next;
ms                583 drivers/scsi/mesh.c 			ms->request_qtail = prev;
ms                585 drivers/scsi/mesh.c 		mesh_start_cmd(ms, cmd);
ms                589 drivers/scsi/mesh.c static void mesh_done(struct mesh_state *ms, int start_next)
ms                592 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms                594 drivers/scsi/mesh.c 	cmd = ms->current_req;
ms                595 drivers/scsi/mesh.c 	ms->current_req = NULL;
ms                598 drivers/scsi/mesh.c 		cmd->result = (ms->stat << 16) | cmd->SCp.Status;
ms                599 drivers/scsi/mesh.c 		if (ms->stat == DID_OK)
ms                603 drivers/scsi/mesh.c 			       cmd->result, ms->data_ptr, scsi_bufflen(cmd));
ms                614 drivers/scsi/mesh.c 		cmd->SCp.this_residual -= ms->data_ptr;
ms                615 drivers/scsi/mesh.c 		mesh_completed(ms, cmd);
ms                618 drivers/scsi/mesh.c 		out_8(&ms->mesh->sequence, SEQ_ENBRESEL);
ms                619 drivers/scsi/mesh.c 		mesh_flush_io(ms->mesh);
ms                621 drivers/scsi/mesh.c 		ms->phase = idle;
ms                622 drivers/scsi/mesh.c 		mesh_start(ms);
ms                626 drivers/scsi/mesh.c static inline void add_sdtr_msg(struct mesh_state *ms)
ms                628 drivers/scsi/mesh.c 	int i = ms->n_msgout;
ms                630 drivers/scsi/mesh.c 	ms->msgout[i] = EXTENDED_MESSAGE;
ms                631 drivers/scsi/mesh.c 	ms->msgout[i+1] = 3;
ms                632 drivers/scsi/mesh.c 	ms->msgout[i+2] = EXTENDED_SDTR;
ms                633 drivers/scsi/mesh.c 	ms->msgout[i+3] = mesh_sync_period/4;
ms                634 drivers/scsi/mesh.c 	ms->msgout[i+4] = (ALLOW_SYNC(ms->conn_tgt)? mesh_sync_offset: 0);
ms                635 drivers/scsi/mesh.c 	ms->n_msgout = i + 5;
ms                638 drivers/scsi/mesh.c static void set_sdtr(struct mesh_state *ms, int period, int offset)
ms                640 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms                641 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                649 drivers/scsi/mesh.c 			       ms->conn_tgt);
ms                658 drivers/scsi/mesh.c 	v = (ms->clk_freq / 5000) * period;
ms                663 drivers/scsi/mesh.c 		tr = (ms->clk_freq + 250000) / 500000;
ms                669 drivers/scsi/mesh.c 		tr = ((ms->clk_freq / (v + 2)) + 199999) / 200000;
ms                676 drivers/scsi/mesh.c 	       ms->conn_tgt, tr/10, tr%10);
ms                679 drivers/scsi/mesh.c static void start_phase(struct mesh_state *ms)
ms                682 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                683 drivers/scsi/mesh.c 	volatile struct dbdma_regs __iomem *md = ms->dma;
ms                684 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd = ms->current_req;
ms                685 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms                687 drivers/scsi/mesh.c 	dlog(ms, "start_phase nmo/exc/fc/seq = %.8x",
ms                688 drivers/scsi/mesh.c 	     MKWORD(ms->n_msgout, mr->exception, mr->fifo_count, mr->sequence));
ms                690 drivers/scsi/mesh.c 	seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
ms                691 drivers/scsi/mesh.c 	switch (ms->msgphase) {
ms                699 drivers/scsi/mesh.c 		ms->n_msgin = 0;
ms                708 drivers/scsi/mesh.c 		if (ms->n_msgout <= 0) {
ms                710 drivers/scsi/mesh.c 			       ms->n_msgout);
ms                711 drivers/scsi/mesh.c 			mesh_dump_regs(ms);
ms                712 drivers/scsi/mesh.c 			ms->msgphase = msg_none;
ms                715 drivers/scsi/mesh.c 		if (ALLOW_DEBUG(ms->conn_tgt)) {
ms                717 drivers/scsi/mesh.c 			       ms->n_msgout);
ms                718 drivers/scsi/mesh.c 			for (i = 0; i < ms->n_msgout; ++i)
ms                719 drivers/scsi/mesh.c 				printk(" %x", ms->msgout[i]);
ms                722 drivers/scsi/mesh.c 		dlog(ms, "msgout msg=%.8x", MKWORD(ms->n_msgout, ms->msgout[0],
ms                723 drivers/scsi/mesh.c 						ms->msgout[1], ms->msgout[2]));
ms                733 drivers/scsi/mesh.c 			dlog(ms, "bus0 was %.2x explicitly asserting ATN", mr->bus_status0);
ms                740 drivers/scsi/mesh.c 			dlog(ms,"hace: after explicit ATN bus0=%.2x",mr->bus_status0);
ms                742 drivers/scsi/mesh.c 		if (ms->n_msgout == 1) {
ms                749 drivers/scsi/mesh.c 			cmd_complete(ms);
ms                751 drivers/scsi/mesh.c 			out_8(&mr->count_lo, ms->n_msgout - 1);
ms                753 drivers/scsi/mesh.c 			for (i = 0; i < ms->n_msgout - 1; ++i)
ms                754 drivers/scsi/mesh.c 				out_8(&mr->fifo, ms->msgout[i]);
ms                760 drivers/scsi/mesh.c 		       ms->msgphase);
ms                763 drivers/scsi/mesh.c 	switch (ms->phase) {
ms                765 drivers/scsi/mesh.c 		out_8(&mr->dest_id, ms->conn_tgt);
ms                785 drivers/scsi/mesh.c 		if (!ms->dma_started) {
ms                786 drivers/scsi/mesh.c 			set_dma_cmds(ms, cmd);
ms                787 drivers/scsi/mesh.c 			out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds));
ms                789 drivers/scsi/mesh.c 			ms->dma_started = 1;
ms                791 drivers/scsi/mesh.c 		nb = ms->dma_count;
ms                794 drivers/scsi/mesh.c 		ms->dma_count -= nb;
ms                795 drivers/scsi/mesh.c 		ms->data_ptr += nb;
ms                811 drivers/scsi/mesh.c 		dlog(ms, "enbresel intr/exc/err/fc=%.8x",
ms                818 drivers/scsi/mesh.c 		       ms->phase);
ms                819 drivers/scsi/mesh.c 		dumpslog(ms);
ms                824 drivers/scsi/mesh.c static inline void get_msgin(struct mesh_state *ms)
ms                826 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                831 drivers/scsi/mesh.c 		i = ms->n_msgin;
ms                832 drivers/scsi/mesh.c 		ms->n_msgin = i + n;
ms                834 drivers/scsi/mesh.c 			ms->msgin[i++] = in_8(&mr->fifo);
ms                838 drivers/scsi/mesh.c static inline int msgin_length(struct mesh_state *ms)
ms                843 drivers/scsi/mesh.c 	if (ms->n_msgin > 0) {
ms                844 drivers/scsi/mesh.c 		b = ms->msgin[0];
ms                847 drivers/scsi/mesh.c 			n = ms->n_msgin < 2? 2: ms->msgin[1] + 2;
ms                856 drivers/scsi/mesh.c static void reselected(struct mesh_state *ms)
ms                858 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                863 drivers/scsi/mesh.c 	switch (ms->phase) {
ms                867 drivers/scsi/mesh.c 		if ((cmd = ms->current_req) != NULL) {
ms                869 drivers/scsi/mesh.c 			cmd->host_scribble = (void *) ms->request_q;
ms                870 drivers/scsi/mesh.c 			if (ms->request_q == NULL)
ms                871 drivers/scsi/mesh.c 				ms->request_qtail = cmd;
ms                872 drivers/scsi/mesh.c 			ms->request_q = cmd;
ms                873 drivers/scsi/mesh.c 			tp = &ms->tgts[cmd->device->id];
ms                878 drivers/scsi/mesh.c 		ms->phase = reselecting;
ms                879 drivers/scsi/mesh.c 		mesh_done(ms, 0);
ms                885 drivers/scsi/mesh.c 		       ms->msgphase, ms->phase, ms->conn_tgt);
ms                886 drivers/scsi/mesh.c 		dumplog(ms, ms->conn_tgt);
ms                887 drivers/scsi/mesh.c 		dumpslog(ms);
ms                890 drivers/scsi/mesh.c 	if (ms->dma_started) {
ms                892 drivers/scsi/mesh.c 		halt_dma(ms);
ms                894 drivers/scsi/mesh.c 	ms->current_req = NULL;
ms                895 drivers/scsi/mesh.c 	ms->phase = dataing;
ms                896 drivers/scsi/mesh.c 	ms->msgphase = msg_in;
ms                897 drivers/scsi/mesh.c 	ms->n_msgout = 0;
ms                898 drivers/scsi/mesh.c 	ms->last_n_msgout = 0;
ms                899 drivers/scsi/mesh.c 	prev = ms->conn_tgt;
ms                913 drivers/scsi/mesh.c 		dlog(ms, "extra resel err/exc/fc = %.6x",
ms                929 drivers/scsi/mesh.c 		ms->conn_tgt = ms->host->this_id;
ms                935 drivers/scsi/mesh.c 		dlog(ms, "reseldata %x", b);
ms                938 drivers/scsi/mesh.c 		if ((b & (1 << t)) != 0 && t != ms->host->this_id)
ms                940 drivers/scsi/mesh.c 	if (b != (1 << t) + (1 << ms->host->this_id)) {
ms                942 drivers/scsi/mesh.c 		ms->conn_tgt = ms->host->this_id;
ms                950 drivers/scsi/mesh.c 	ms->conn_tgt = t;
ms                951 drivers/scsi/mesh.c 	tp = &ms->tgts[t];
ms                958 drivers/scsi/mesh.c 	ms->current_req = tp->current_req;
ms                963 drivers/scsi/mesh.c 	ms->data_ptr = tp->saved_ptr;
ms                964 drivers/scsi/mesh.c 	dlog(ms, "resel prev tgt=%d", prev);
ms                965 drivers/scsi/mesh.c 	dlog(ms, "resel err/exc=%.4x", MKWORD(0, 0, mr->error, mr->exception));
ms                966 drivers/scsi/mesh.c 	start_phase(ms);
ms                970 drivers/scsi/mesh.c 	dumplog(ms, ms->conn_tgt);
ms                971 drivers/scsi/mesh.c 	dumpslog(ms);
ms                972 drivers/scsi/mesh.c 	ms->data_ptr = 0;
ms                973 drivers/scsi/mesh.c 	ms->aborting = 1;
ms                974 drivers/scsi/mesh.c 	start_phase(ms);
ms                977 drivers/scsi/mesh.c static void do_abort(struct mesh_state *ms)
ms                979 drivers/scsi/mesh.c 	ms->msgout[0] = ABORT;
ms                980 drivers/scsi/mesh.c 	ms->n_msgout = 1;
ms                981 drivers/scsi/mesh.c 	ms->aborting = 1;
ms                982 drivers/scsi/mesh.c 	ms->stat = DID_ABORT;
ms                983 drivers/scsi/mesh.c 	dlog(ms, "abort", 0);
ms                986 drivers/scsi/mesh.c static void handle_reset(struct mesh_state *ms)
ms                991 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms                994 drivers/scsi/mesh.c 		tp = &ms->tgts[tgt];
ms                998 drivers/scsi/mesh.c 			mesh_completed(ms, cmd);
ms               1000 drivers/scsi/mesh.c 		ms->tgts[tgt].sdtr_state = do_sdtr;
ms               1001 drivers/scsi/mesh.c 		ms->tgts[tgt].sync_params = ASYNC_PARAMS;
ms               1003 drivers/scsi/mesh.c 	ms->current_req = NULL;
ms               1004 drivers/scsi/mesh.c 	while ((cmd = ms->request_q) != NULL) {
ms               1005 drivers/scsi/mesh.c 		ms->request_q = (struct scsi_cmnd *) cmd->host_scribble;
ms               1007 drivers/scsi/mesh.c 		mesh_completed(ms, cmd);
ms               1009 drivers/scsi/mesh.c 	ms->phase = idle;
ms               1010 drivers/scsi/mesh.c 	ms->msgphase = msg_none;
ms               1022 drivers/scsi/mesh.c 	struct mesh_state *ms = dev_id;
ms               1023 drivers/scsi/mesh.c 	struct Scsi_Host *dev = ms->host;
ms               1026 drivers/scsi/mesh.c 	mesh_interrupt(ms);
ms               1031 drivers/scsi/mesh.c static void handle_error(struct mesh_state *ms)
ms               1034 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1039 drivers/scsi/mesh.c 	dlog(ms, "error err/exc/fc/cl=%.8x",
ms               1048 drivers/scsi/mesh.c 		handle_reset(ms);
ms               1055 drivers/scsi/mesh.c 			reselected(ms);
ms               1058 drivers/scsi/mesh.c 		if (!ms->aborting) {
ms               1060 drivers/scsi/mesh.c 			       ms->conn_tgt);
ms               1061 drivers/scsi/mesh.c 			dumplog(ms, ms->conn_tgt);
ms               1062 drivers/scsi/mesh.c 			dumpslog(ms);
ms               1065 drivers/scsi/mesh.c 		ms->stat = DID_ABORT;
ms               1066 drivers/scsi/mesh.c 		mesh_done(ms, 1);
ms               1070 drivers/scsi/mesh.c 		if (ms->msgphase == msg_in) {
ms               1072 drivers/scsi/mesh.c 			       ms->conn_tgt);
ms               1073 drivers/scsi/mesh.c 			ms->msgout[0] = MSG_PARITY_ERROR;
ms               1074 drivers/scsi/mesh.c 			ms->n_msgout = 1;
ms               1075 drivers/scsi/mesh.c 			ms->msgphase = msg_in_bad;
ms               1076 drivers/scsi/mesh.c 			cmd_complete(ms);
ms               1079 drivers/scsi/mesh.c 		if (ms->stat == DID_OK) {
ms               1081 drivers/scsi/mesh.c 			       ms->conn_tgt);
ms               1082 drivers/scsi/mesh.c 			ms->stat = DID_PARITY;
ms               1086 drivers/scsi/mesh.c 			cmd_complete(ms);
ms               1099 drivers/scsi/mesh.c 			reselected(ms);
ms               1105 drivers/scsi/mesh.c 			phase_mismatch(ms);
ms               1113 drivers/scsi/mesh.c 	mesh_dump_regs(ms);
ms               1114 drivers/scsi/mesh.c 	dumplog(ms, ms->conn_tgt);
ms               1115 drivers/scsi/mesh.c 	if (ms->phase > selecting && (in_8(&mr->bus_status1) & BS1_BSY)) {
ms               1117 drivers/scsi/mesh.c 		do_abort(ms);
ms               1118 drivers/scsi/mesh.c 		phase_mismatch(ms);
ms               1121 drivers/scsi/mesh.c 	ms->stat = DID_ERROR;
ms               1122 drivers/scsi/mesh.c 	mesh_done(ms, 1);
ms               1125 drivers/scsi/mesh.c static void handle_exception(struct mesh_state *ms)
ms               1128 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1135 drivers/scsi/mesh.c 		reselected(ms);
ms               1138 drivers/scsi/mesh.c 		ms->stat = DID_BUS_BUSY;
ms               1139 drivers/scsi/mesh.c 		mesh_done(ms, 1);
ms               1142 drivers/scsi/mesh.c 		ms->stat = DID_BAD_TARGET;
ms               1143 drivers/scsi/mesh.c 		mesh_done(ms, 1);
ms               1147 drivers/scsi/mesh.c 		phase_mismatch(ms);
ms               1150 drivers/scsi/mesh.c 		mesh_dump_regs(ms);
ms               1151 drivers/scsi/mesh.c 		dumplog(ms, ms->conn_tgt);
ms               1152 drivers/scsi/mesh.c 		do_abort(ms);
ms               1153 drivers/scsi/mesh.c 		phase_mismatch(ms);
ms               1157 drivers/scsi/mesh.c static void handle_msgin(struct mesh_state *ms)
ms               1160 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd = ms->current_req;
ms               1161 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms               1163 drivers/scsi/mesh.c 	if (ms->n_msgin == 0)
ms               1165 drivers/scsi/mesh.c 	code = ms->msgin[0];
ms               1166 drivers/scsi/mesh.c 	if (ALLOW_DEBUG(ms->conn_tgt)) {
ms               1167 drivers/scsi/mesh.c 		printk(KERN_DEBUG "got %d message bytes:", ms->n_msgin);
ms               1168 drivers/scsi/mesh.c 		for (i = 0; i < ms->n_msgin; ++i)
ms               1169 drivers/scsi/mesh.c 			printk(" %x", ms->msgin[i]);
ms               1172 drivers/scsi/mesh.c 	dlog(ms, "msgin msg=%.8x",
ms               1173 drivers/scsi/mesh.c 	     MKWORD(ms->n_msgin, code, ms->msgin[1], ms->msgin[2]));
ms               1175 drivers/scsi/mesh.c 	ms->expect_reply = 0;
ms               1176 drivers/scsi/mesh.c 	ms->n_msgout = 0;
ms               1177 drivers/scsi/mesh.c 	if (ms->n_msgin < msgin_length(ms))
ms               1185 drivers/scsi/mesh.c 		switch (ms->msgin[2]) {
ms               1187 drivers/scsi/mesh.c 			ms->data_ptr += (ms->msgin[3] << 24) + ms->msgin[6]
ms               1188 drivers/scsi/mesh.c 				+ (ms->msgin[4] << 16) + (ms->msgin[5] << 8);
ms               1193 drivers/scsi/mesh.c 				add_sdtr_msg(ms);
ms               1196 drivers/scsi/mesh.c 				if (ms->msgout[3] < ms->msgin[3])
ms               1197 drivers/scsi/mesh.c 					ms->msgout[3] = ms->msgin[3];
ms               1198 drivers/scsi/mesh.c 				if (ms->msgout[4] > ms->msgin[4])
ms               1199 drivers/scsi/mesh.c 					ms->msgout[4] = ms->msgin[4];
ms               1200 drivers/scsi/mesh.c 				set_sdtr(ms, ms->msgout[3], ms->msgout[4]);
ms               1201 drivers/scsi/mesh.c 				ms->msgphase = msg_out;
ms               1203 drivers/scsi/mesh.c 				set_sdtr(ms, ms->msgin[3], ms->msgin[4]);
ms               1211 drivers/scsi/mesh.c 		tp->saved_ptr = ms->data_ptr;
ms               1214 drivers/scsi/mesh.c 		ms->data_ptr = tp->saved_ptr;
ms               1217 drivers/scsi/mesh.c 		ms->phase = disconnecting;
ms               1223 drivers/scsi/mesh.c 			set_sdtr(ms, 0, 0);
ms               1230 drivers/scsi/mesh.c 				do_abort(ms);
ms               1231 drivers/scsi/mesh.c 				ms->msgphase = msg_out;
ms               1236 drivers/scsi/mesh.c 				       cmd->device->lun, ms->conn_tgt);
ms               1246 drivers/scsi/mesh.c 	       ms->conn_tgt);
ms               1247 drivers/scsi/mesh.c 	for (i = 0; i < ms->n_msgin; ++i)
ms               1248 drivers/scsi/mesh.c 		printk(" %x", ms->msgin[i]);
ms               1250 drivers/scsi/mesh.c 	ms->msgout[0] = MESSAGE_REJECT;
ms               1251 drivers/scsi/mesh.c 	ms->n_msgout = 1;
ms               1252 drivers/scsi/mesh.c 	ms->msgphase = msg_out;
ms               1258 drivers/scsi/mesh.c static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd)
ms               1264 drivers/scsi/mesh.c 	dma_cmd = ms->tgts[ms->conn_tgt].data_goes_out?
ms               1266 drivers/scsi/mesh.c 	dcmds = ms->dma_cmds;
ms               1278 drivers/scsi/mesh.c 			off = ms->data_ptr;
ms               1316 drivers/scsi/mesh.c 	ms->dma_count = dtot;
ms               1319 drivers/scsi/mesh.c static void halt_dma(struct mesh_state *ms)
ms               1321 drivers/scsi/mesh.c 	volatile struct dbdma_regs __iomem *md = ms->dma;
ms               1322 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1323 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd = ms->current_req;
ms               1326 drivers/scsi/mesh.c 	if (!ms->tgts[ms->conn_tgt].data_goes_out) {
ms               1337 drivers/scsi/mesh.c 	dlog(ms, "halt_dma fc/count=%.6x",
ms               1339 drivers/scsi/mesh.c 	if (ms->tgts[ms->conn_tgt].data_goes_out)
ms               1343 drivers/scsi/mesh.c 	ms->data_ptr -= nb;
ms               1344 drivers/scsi/mesh.c 	dlog(ms, "data_ptr %x", ms->data_ptr);
ms               1345 drivers/scsi/mesh.c 	if (ms->data_ptr < 0) {
ms               1347 drivers/scsi/mesh.c 		       ms->data_ptr, nb, ms);
ms               1348 drivers/scsi/mesh.c 		ms->data_ptr = 0;
ms               1350 drivers/scsi/mesh.c 		dumplog(ms, ms->conn_tgt);
ms               1351 drivers/scsi/mesh.c 		dumpslog(ms);
ms               1354 drivers/scsi/mesh.c 		   ms->data_ptr > scsi_bufflen(cmd)) {
ms               1357 drivers/scsi/mesh.c 		       ms->conn_tgt, ms->data_ptr, scsi_bufflen(cmd),
ms               1358 drivers/scsi/mesh.c 		       ms->tgts[ms->conn_tgt].data_goes_out);
ms               1361 drivers/scsi/mesh.c 	ms->dma_started = 0;
ms               1364 drivers/scsi/mesh.c static void phase_mismatch(struct mesh_state *ms)
ms               1366 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1369 drivers/scsi/mesh.c 	dlog(ms, "phasemm ch/cl/seq/fc=%.8x",
ms               1372 drivers/scsi/mesh.c 	if (ms->msgphase == msg_out_xxx && phase == BP_MSGOUT) {
ms               1378 drivers/scsi/mesh.c 		out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
ms               1379 drivers/scsi/mesh.c 		ms->msgphase = msg_out_last;
ms               1383 drivers/scsi/mesh.c 	if (ms->msgphase == msg_in) {
ms               1384 drivers/scsi/mesh.c 		get_msgin(ms);
ms               1385 drivers/scsi/mesh.c 		if (ms->n_msgin)
ms               1386 drivers/scsi/mesh.c 			handle_msgin(ms);
ms               1389 drivers/scsi/mesh.c 	if (ms->dma_started)
ms               1390 drivers/scsi/mesh.c 		halt_dma(ms);
ms               1397 drivers/scsi/mesh.c 	ms->msgphase = msg_none;
ms               1400 drivers/scsi/mesh.c 		ms->tgts[ms->conn_tgt].data_goes_out = 0;
ms               1401 drivers/scsi/mesh.c 		ms->phase = dataing;
ms               1404 drivers/scsi/mesh.c 		ms->tgts[ms->conn_tgt].data_goes_out = 1;
ms               1405 drivers/scsi/mesh.c 		ms->phase = dataing;
ms               1408 drivers/scsi/mesh.c 		ms->phase = commanding;
ms               1411 drivers/scsi/mesh.c 		ms->phase = statusing;
ms               1414 drivers/scsi/mesh.c 		ms->msgphase = msg_in;
ms               1415 drivers/scsi/mesh.c 		ms->n_msgin = 0;
ms               1418 drivers/scsi/mesh.c 		ms->msgphase = msg_out;
ms               1419 drivers/scsi/mesh.c 		if (ms->n_msgout == 0) {
ms               1420 drivers/scsi/mesh.c 			if (ms->aborting) {
ms               1421 drivers/scsi/mesh.c 				do_abort(ms);
ms               1423 drivers/scsi/mesh.c 				if (ms->last_n_msgout == 0) {
ms               1426 drivers/scsi/mesh.c 					ms->msgout[0] = NOP;
ms               1427 drivers/scsi/mesh.c 					ms->last_n_msgout = 1;
ms               1429 drivers/scsi/mesh.c 				ms->n_msgout = ms->last_n_msgout;
ms               1435 drivers/scsi/mesh.c 		ms->stat = DID_ERROR;
ms               1436 drivers/scsi/mesh.c 		mesh_done(ms, 1);
ms               1440 drivers/scsi/mesh.c 	start_phase(ms);
ms               1443 drivers/scsi/mesh.c static void cmd_complete(struct mesh_state *ms)
ms               1445 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1446 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd = ms->current_req;
ms               1447 drivers/scsi/mesh.c 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
ms               1450 drivers/scsi/mesh.c 	dlog(ms, "cmd_complete fc=%x", mr->fifo_count);
ms               1451 drivers/scsi/mesh.c 	seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
ms               1452 drivers/scsi/mesh.c 	switch (ms->msgphase) {
ms               1455 drivers/scsi/mesh.c 		ms->n_msgin = 0;
ms               1456 drivers/scsi/mesh.c 		ms->msgphase = msg_in;
ms               1461 drivers/scsi/mesh.c 		get_msgin(ms);
ms               1462 drivers/scsi/mesh.c 		n = msgin_length(ms);
ms               1463 drivers/scsi/mesh.c 		if (ms->n_msgin < n) {
ms               1464 drivers/scsi/mesh.c 			out_8(&mr->count_lo, n - ms->n_msgin);
ms               1467 drivers/scsi/mesh.c 			ms->msgphase = msg_none;
ms               1468 drivers/scsi/mesh.c 			handle_msgin(ms);
ms               1469 drivers/scsi/mesh.c 			start_phase(ms);
ms               1498 drivers/scsi/mesh.c 		dlog(ms, "last_mbyte err/exc/fc/cl=%.8x",
ms               1503 drivers/scsi/mesh.c 			ms->last_n_msgout = ms->n_msgout;
ms               1504 drivers/scsi/mesh.c 			ms->n_msgout = 0;
ms               1508 drivers/scsi/mesh.c 				handle_error(ms);
ms               1517 drivers/scsi/mesh.c 			handle_exception(ms);
ms               1524 drivers/scsi/mesh.c 			out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
ms               1525 drivers/scsi/mesh.c 			ms->msgphase = msg_out_last;
ms               1528 drivers/scsi/mesh.c 			ms->msgphase = msg_out_xxx;
ms               1533 drivers/scsi/mesh.c 		ms->last_n_msgout = ms->n_msgout;
ms               1534 drivers/scsi/mesh.c 		ms->n_msgout = 0;
ms               1535 drivers/scsi/mesh.c 		ms->msgphase = ms->expect_reply? msg_in: msg_none;
ms               1536 drivers/scsi/mesh.c 		start_phase(ms);
ms               1540 drivers/scsi/mesh.c 		switch (ms->phase) {
ms               1543 drivers/scsi/mesh.c 			dumpslog(ms);
ms               1546 drivers/scsi/mesh.c 			dlog(ms, "Selecting phase at command completion",0);
ms               1547 drivers/scsi/mesh.c 			ms->msgout[0] = IDENTIFY(ALLOW_RESEL(ms->conn_tgt),
ms               1549 drivers/scsi/mesh.c 			ms->n_msgout = 1;
ms               1550 drivers/scsi/mesh.c 			ms->expect_reply = 0;
ms               1551 drivers/scsi/mesh.c 			if (ms->aborting) {
ms               1552 drivers/scsi/mesh.c 				ms->msgout[0] = ABORT;
ms               1553 drivers/scsi/mesh.c 				ms->n_msgout++;
ms               1556 drivers/scsi/mesh.c 				add_sdtr_msg(ms);
ms               1557 drivers/scsi/mesh.c 				ms->expect_reply = 1;
ms               1560 drivers/scsi/mesh.c 			ms->msgphase = msg_out;
ms               1571 drivers/scsi/mesh.c 					dlog(ms, "impatient for req", ms->n_msgout);
ms               1572 drivers/scsi/mesh.c 					ms->msgphase = msg_none;
ms               1579 drivers/scsi/mesh.c 			if (ms->dma_count != 0) {
ms               1580 drivers/scsi/mesh.c 				start_phase(ms);
ms               1596 drivers/scsi/mesh.c 			halt_dma(ms);
ms               1605 drivers/scsi/mesh.c 			ms->msgphase = msg_in;
ms               1608 drivers/scsi/mesh.c 			mesh_done(ms, 1);
ms               1611 drivers/scsi/mesh.c 			ms->current_req = NULL;
ms               1612 drivers/scsi/mesh.c 			ms->phase = idle;
ms               1613 drivers/scsi/mesh.c 			mesh_start(ms);
ms               1618 drivers/scsi/mesh.c 		++ms->phase;
ms               1619 drivers/scsi/mesh.c 		start_phase(ms);
ms               1631 drivers/scsi/mesh.c 	struct mesh_state *ms;
ms               1636 drivers/scsi/mesh.c 	ms = (struct mesh_state *) cmd->device->host->hostdata;
ms               1638 drivers/scsi/mesh.c 	if (ms->request_q == NULL)
ms               1639 drivers/scsi/mesh.c 		ms->request_q = cmd;
ms               1641 drivers/scsi/mesh.c 		ms->request_qtail->host_scribble = (void *) cmd;
ms               1642 drivers/scsi/mesh.c 	ms->request_qtail = cmd;
ms               1644 drivers/scsi/mesh.c 	if (ms->phase == idle)
ms               1645 drivers/scsi/mesh.c 		mesh_start(ms);
ms               1657 drivers/scsi/mesh.c static void mesh_interrupt(struct mesh_state *ms)
ms               1659 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1663 drivers/scsi/mesh.c 	if (ALLOW_DEBUG(ms->conn_tgt))
ms               1667 drivers/scsi/mesh.c 		       ms->phase, ms->msgphase);
ms               1670 drivers/scsi/mesh.c 		dlog(ms, "interrupt intr/err/exc/seq=%.8x", 
ms               1673 drivers/scsi/mesh.c 			handle_error(ms);
ms               1675 drivers/scsi/mesh.c 			handle_exception(ms);
ms               1678 drivers/scsi/mesh.c 			cmd_complete(ms);
ms               1689 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
ms               1692 drivers/scsi/mesh.c 	mesh_dump_regs(ms);
ms               1693 drivers/scsi/mesh.c 	dumplog(ms, cmd->device->id);
ms               1694 drivers/scsi/mesh.c 	dumpslog(ms);
ms               1706 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
ms               1707 drivers/scsi/mesh.c 	volatile struct mesh_regs __iomem *mr = ms->mesh;
ms               1708 drivers/scsi/mesh.c 	volatile struct dbdma_regs __iomem *md = ms->dma;
ms               1713 drivers/scsi/mesh.c 	spin_lock_irqsave(ms->host->host_lock, flags);
ms               1723 drivers/scsi/mesh.c 	out_8(&mr->source_id, ms->host->this_id);
ms               1734 drivers/scsi/mesh.c 	handle_reset(ms);
ms               1736 drivers/scsi/mesh.c 	spin_unlock_irqrestore(ms->host->host_lock, flags);
ms               1740 drivers/scsi/mesh.c static void set_mesh_power(struct mesh_state *ms, int state)
ms               1745 drivers/scsi/mesh.c 		pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 1);
ms               1748 drivers/scsi/mesh.c 		pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 0);
ms               1757 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
ms               1768 drivers/scsi/mesh.c 	if (ms->phase == sleeping)
ms               1771 drivers/scsi/mesh.c 	scsi_block_requests(ms->host);
ms               1772 drivers/scsi/mesh.c 	spin_lock_irqsave(ms->host->host_lock, flags);
ms               1773 drivers/scsi/mesh.c 	while(ms->phase != idle) {
ms               1774 drivers/scsi/mesh.c 		spin_unlock_irqrestore(ms->host->host_lock, flags);
ms               1776 drivers/scsi/mesh.c 		spin_lock_irqsave(ms->host->host_lock, flags);
ms               1778 drivers/scsi/mesh.c 	ms->phase = sleeping;
ms               1779 drivers/scsi/mesh.c 	spin_unlock_irqrestore(ms->host->host_lock, flags);
ms               1780 drivers/scsi/mesh.c 	disable_irq(ms->meshintr);
ms               1781 drivers/scsi/mesh.c 	set_mesh_power(ms, 0);
ms               1788 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
ms               1791 drivers/scsi/mesh.c 	if (ms->phase != sleeping)
ms               1794 drivers/scsi/mesh.c 	set_mesh_power(ms, 1);
ms               1795 drivers/scsi/mesh.c 	mesh_init(ms);
ms               1796 drivers/scsi/mesh.c 	spin_lock_irqsave(ms->host->host_lock, flags);
ms               1797 drivers/scsi/mesh.c 	mesh_start(ms);
ms               1798 drivers/scsi/mesh.c 	spin_unlock_irqrestore(ms->host->host_lock, flags);
ms               1799 drivers/scsi/mesh.c 	enable_irq(ms->meshintr);
ms               1800 drivers/scsi/mesh.c 	scsi_unblock_requests(ms->host);
ms               1814 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
ms               1819 drivers/scsi/mesh.c 	spin_lock_irqsave(ms->host->host_lock, flags);
ms               1820 drivers/scsi/mesh.c        	mr = ms->mesh;
ms               1827 drivers/scsi/mesh.c 	spin_unlock_irqrestore(ms->host->host_lock, flags);
ms               1851 drivers/scsi/mesh.c 	struct mesh_state *ms;
ms               1890 drivers/scsi/mesh.c        	ms = (struct mesh_state *) mesh_host->hostdata;
ms               1891 drivers/scsi/mesh.c 	macio_set_drvdata(mdev, ms);
ms               1892 drivers/scsi/mesh.c 	ms->host = mesh_host;
ms               1893 drivers/scsi/mesh.c 	ms->mdev = mdev;
ms               1894 drivers/scsi/mesh.c 	ms->pdev = pdev;
ms               1896 drivers/scsi/mesh.c 	ms->mesh = ioremap(macio_resource_start(mdev, 0), 0x1000);
ms               1897 drivers/scsi/mesh.c 	if (ms->mesh == NULL) {
ms               1901 drivers/scsi/mesh.c 	ms->dma = ioremap(macio_resource_start(mdev, 1), 0x1000);
ms               1902 drivers/scsi/mesh.c 	if (ms->dma == NULL) {
ms               1904 drivers/scsi/mesh.c 		iounmap(ms->mesh);
ms               1908 drivers/scsi/mesh.c        	ms->meshintr = macio_irq(mdev, 0);
ms               1909 drivers/scsi/mesh.c        	ms->dmaintr = macio_irq(mdev, 1);
ms               1914 drivers/scsi/mesh.c 	ms->dma_cmd_size = (mesh_host->sg_tablesize + 2) * sizeof(struct dbdma_cmd);
ms               1920 drivers/scsi/mesh.c 					   ms->dma_cmd_size, &dma_cmd_bus,
ms               1927 drivers/scsi/mesh.c 	ms->dma_cmds = (struct dbdma_cmd *) DBDMA_ALIGN(dma_cmd_space);
ms               1928 drivers/scsi/mesh.c        	ms->dma_cmd_space = dma_cmd_space;
ms               1929 drivers/scsi/mesh.c 	ms->dma_cmd_bus = dma_cmd_bus + ((unsigned long)ms->dma_cmds)
ms               1931 drivers/scsi/mesh.c 	ms->current_req = NULL;
ms               1933 drivers/scsi/mesh.c 	       	ms->tgts[tgt].sdtr_state = do_sdtr;
ms               1934 drivers/scsi/mesh.c 	       	ms->tgts[tgt].sync_params = ASYNC_PARAMS;
ms               1935 drivers/scsi/mesh.c 	       	ms->tgts[tgt].current_req = NULL;
ms               1939 drivers/scsi/mesh.c        		ms->clk_freq = *cfp;
ms               1942 drivers/scsi/mesh.c 	       	ms->clk_freq = 50000000;
ms               1948 drivers/scsi/mesh.c 	minper = 1000000000 / (ms->clk_freq / 5); /* ns */
ms               1953 drivers/scsi/mesh.c 	set_mesh_power(ms, 1);
ms               1956 drivers/scsi/mesh.c        	mesh_init(ms);
ms               1959 drivers/scsi/mesh.c        	if (request_irq(ms->meshintr, do_mesh_interrupt, 0, "MESH", ms)) {
ms               1960 drivers/scsi/mesh.c 	       	printk(KERN_ERR "MESH: can't get irq %d\n", ms->meshintr);
ms               1972 drivers/scsi/mesh.c 	free_irq(ms->meshintr, ms);
ms               1978 drivers/scsi/mesh.c 	set_mesh_power(ms, 0);
ms               1979 drivers/scsi/mesh.c 	dma_free_coherent(&macio_get_pci_dev(mdev)->dev, ms->dma_cmd_size,
ms               1980 drivers/scsi/mesh.c 			    ms->dma_cmd_space, ms->dma_cmd_bus);
ms               1982 drivers/scsi/mesh.c 	iounmap(ms->dma);
ms               1983 drivers/scsi/mesh.c 	iounmap(ms->mesh);
ms               1994 drivers/scsi/mesh.c 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
ms               1995 drivers/scsi/mesh.c 	struct Scsi_Host *mesh_host = ms->host;
ms               1999 drivers/scsi/mesh.c 	free_irq(ms->meshintr, ms);
ms               2005 drivers/scsi/mesh.c 	set_mesh_power(ms, 0);
ms               2008 drivers/scsi/mesh.c 	iounmap(ms->mesh);
ms               2009 drivers/scsi/mesh.c        	iounmap(ms->dma);
ms               2012 drivers/scsi/mesh.c 	dma_free_coherent(&macio_get_pci_dev(mdev)->dev, ms->dma_cmd_size,
ms               2013 drivers/scsi/mesh.c 			    ms->dma_cmd_space, ms->dma_cmd_bus);
ms               7843 drivers/scsi/ncr53c8xx.c 	unsigned ms = 0;
ms               7869 drivers/scsi/ncr53c8xx.c 	while (!(INW(nc_sist) & GEN) && ms++ < 100000) {
ms               7882 drivers/scsi/ncr53c8xx.c 		printk ("%s: Delay (GEN=%d): %u msec\n", ncr_name(np), gen, ms);
ms               7886 drivers/scsi/ncr53c8xx.c 	return ms ? ((1 << gen) * 4340) / ms : 0;
ms                872 drivers/scsi/sg.c 	unsigned int ms;
ms                887 drivers/scsi/sg.c 			ms = jiffies_to_msecs(jiffies);
ms                889 drivers/scsi/sg.c 				(ms > srp->header.duration) ?
ms                890 drivers/scsi/sg.c 				(ms - srp->header.duration) : 0;
ms               1327 drivers/scsi/sg.c 	unsigned int ms;
ms               1350 drivers/scsi/sg.c 	ms = jiffies_to_msecs(jiffies);
ms               1351 drivers/scsi/sg.c 	srp->header.duration = (ms > srp->header.duration) ?
ms               1352 drivers/scsi/sg.c 				(ms - srp->header.duration) : 0;
ms               2524 drivers/scsi/sg.c 	unsigned int ms;
ms               2565 drivers/scsi/sg.c 				ms = jiffies_to_msecs(jiffies);
ms               2569 drivers/scsi/sg.c 					(ms > hp->duration ? ms - hp->duration : 0));
ms                291 drivers/scsi/sym53c8xx_2/sym_hipd.c 	unsigned int ms = 0;
ms                325 drivers/scsi/sym53c8xx_2/sym_hipd.c 	while (!(INW(np, nc_sist) & GEN) && ms++ < 100000)
ms                345 drivers/scsi/sym53c8xx_2/sym_hipd.c 	f = ms ? ((1 << gen) * (4340*4)) / ms : 0;
ms                356 drivers/scsi/sym53c8xx_2/sym_hipd.c 			sym_name(np), gen, ms/4, f);
ms                151 drivers/slimbus/messaging.c 		unsigned long ms = txn->rl + HZ;
ms                154 drivers/slimbus/messaging.c 						      msecs_to_jiffies(ms));
ms                331 drivers/slimbus/qcom-ctrl.c 	unsigned long ms = txn->rl + HZ;
ms                377 drivers/slimbus/qcom-ctrl.c 	timeout = wait_for_completion_timeout(&done, msecs_to_jiffies(ms));
ms                 82 drivers/spi/spi-mpc52xx.c 	int (*state)(int irq, struct mpc52xx_spi *ms, u8 status, u8 data);
ms                 95 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_chipsel(struct mpc52xx_spi *ms, int value)
ms                 99 drivers/spi/spi-mpc52xx.c 	if (ms->gpio_cs_count > 0) {
ms                100 drivers/spi/spi-mpc52xx.c 		cs = ms->message->spi->chip_select;
ms                101 drivers/spi/spi-mpc52xx.c 		gpio_set_value(ms->gpio_cs[cs], value ? 0 : 1);
ms                103 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_PORTDATA, value ? 0 : 0x08);
ms                111 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_start_transfer(struct mpc52xx_spi *ms)
ms                113 drivers/spi/spi-mpc52xx.c 	ms->rx_buf = ms->transfer->rx_buf;
ms                114 drivers/spi/spi-mpc52xx.c 	ms->tx_buf = ms->transfer->tx_buf;
ms                115 drivers/spi/spi-mpc52xx.c 	ms->len = ms->transfer->len;
ms                118 drivers/spi/spi-mpc52xx.c 	if (ms->cs_change)
ms                119 drivers/spi/spi-mpc52xx.c 		mpc52xx_spi_chipsel(ms, 1);
ms                120 drivers/spi/spi-mpc52xx.c 	ms->cs_change = ms->transfer->cs_change;
ms                123 drivers/spi/spi-mpc52xx.c 	ms->wcol_tx_timestamp = get_tbl();
ms                124 drivers/spi/spi-mpc52xx.c 	if (ms->tx_buf)
ms                125 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_DATA, *ms->tx_buf++);
ms                127 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_DATA, 0);
ms                131 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms,
ms                133 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms,
ms                143 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsmstate_idle(int irq, struct mpc52xx_spi *ms, u8 status, u8 data)
ms                150 drivers/spi/spi-mpc52xx.c 		dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n",
ms                154 drivers/spi/spi-mpc52xx.c 	if (list_empty(&ms->queue))
ms                158 drivers/spi/spi-mpc52xx.c 	ms->message = list_first_entry(&ms->queue, struct spi_message, queue);
ms                159 drivers/spi/spi-mpc52xx.c 	list_del_init(&ms->message->queue);
ms                163 drivers/spi/spi-mpc52xx.c 	spi = ms->message->spi;
ms                170 drivers/spi/spi-mpc52xx.c 	out_8(ms->regs + SPI_CTRL1, ctrl1);
ms                175 drivers/spi/spi-mpc52xx.c 	sppr = ((ms->ipb_freq / ms->message->spi->max_speed_hz) + 1) >> 1;
ms                189 drivers/spi/spi-mpc52xx.c 	out_8(ms->regs + SPI_BRR, sppr << 4 | spr); /* Set speed */
ms                191 drivers/spi/spi-mpc52xx.c 	ms->cs_change = 1;
ms                192 drivers/spi/spi-mpc52xx.c 	ms->transfer = container_of(ms->message->transfers.next,
ms                195 drivers/spi/spi-mpc52xx.c 	mpc52xx_spi_start_transfer(ms);
ms                196 drivers/spi/spi-mpc52xx.c 	ms->state = mpc52xx_spi_fsmstate_transfer;
ms                209 drivers/spi/spi-mpc52xx.c static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms,
ms                213 drivers/spi/spi-mpc52xx.c 		return ms->irq0 ? FSM_STOP : FSM_POLL;
ms                223 drivers/spi/spi-mpc52xx.c 		ms->wcol_count++;
ms                224 drivers/spi/spi-mpc52xx.c 		ms->wcol_ticks += get_tbl() - ms->wcol_tx_timestamp;
ms                225 drivers/spi/spi-mpc52xx.c 		ms->wcol_tx_timestamp = get_tbl();
ms                227 drivers/spi/spi-mpc52xx.c 		if (ms->tx_buf)
ms                228 drivers/spi/spi-mpc52xx.c 			data = *(ms->tx_buf - 1);
ms                229 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_DATA, data); /* try again */
ms                232 drivers/spi/spi-mpc52xx.c 		ms->modf_count++;
ms                233 drivers/spi/spi-mpc52xx.c 		dev_err(&ms->master->dev, "mode fault\n");
ms                234 drivers/spi/spi-mpc52xx.c 		mpc52xx_spi_chipsel(ms, 0);
ms                235 drivers/spi/spi-mpc52xx.c 		ms->message->status = -EIO;
ms                236 drivers/spi/spi-mpc52xx.c 		if (ms->message->complete)
ms                237 drivers/spi/spi-mpc52xx.c 			ms->message->complete(ms->message->context);
ms                238 drivers/spi/spi-mpc52xx.c 		ms->state = mpc52xx_spi_fsmstate_idle;
ms                243 drivers/spi/spi-mpc52xx.c 	ms->byte_count++;
ms                244 drivers/spi/spi-mpc52xx.c 	if (ms->rx_buf)
ms                245 drivers/spi/spi-mpc52xx.c 		*ms->rx_buf++ = data;
ms                248 drivers/spi/spi-mpc52xx.c 	ms->len--;
ms                249 drivers/spi/spi-mpc52xx.c 	if (ms->len == 0) {
ms                250 drivers/spi/spi-mpc52xx.c 		ms->timestamp = get_tbl();
ms                251 drivers/spi/spi-mpc52xx.c 		ms->timestamp += ms->transfer->delay_usecs * tb_ticks_per_usec;
ms                252 drivers/spi/spi-mpc52xx.c 		ms->state = mpc52xx_spi_fsmstate_wait;
ms                257 drivers/spi/spi-mpc52xx.c 	ms->wcol_tx_timestamp = get_tbl();
ms                258 drivers/spi/spi-mpc52xx.c 	if (ms->tx_buf)
ms                259 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_DATA, *ms->tx_buf++);
ms                261 drivers/spi/spi-mpc52xx.c 		out_8(ms->regs + SPI_DATA, 0);
ms                273 drivers/spi/spi-mpc52xx.c mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms, u8 status, u8 data)
ms                276 drivers/spi/spi-mpc52xx.c 		dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n",
ms                279 drivers/spi/spi-mpc52xx.c 	if (((int)get_tbl()) - ms->timestamp < 0)
ms                282 drivers/spi/spi-mpc52xx.c 	ms->message->actual_length += ms->transfer->len;
ms                287 drivers/spi/spi-mpc52xx.c 	if (ms->transfer->transfer_list.next == &ms->message->transfers) {
ms                288 drivers/spi/spi-mpc52xx.c 		ms->msg_count++;
ms                289 drivers/spi/spi-mpc52xx.c 		mpc52xx_spi_chipsel(ms, 0);
ms                290 drivers/spi/spi-mpc52xx.c 		ms->message->status = 0;
ms                291 drivers/spi/spi-mpc52xx.c 		if (ms->message->complete)
ms                292 drivers/spi/spi-mpc52xx.c 			ms->message->complete(ms->message->context);
ms                293 drivers/spi/spi-mpc52xx.c 		ms->state = mpc52xx_spi_fsmstate_idle;
ms                299 drivers/spi/spi-mpc52xx.c 	if (ms->cs_change)
ms                300 drivers/spi/spi-mpc52xx.c 		mpc52xx_spi_chipsel(ms, 0);
ms                302 drivers/spi/spi-mpc52xx.c 	ms->transfer = container_of(ms->transfer->transfer_list.next,
ms                304 drivers/spi/spi-mpc52xx.c 	mpc52xx_spi_start_transfer(ms);
ms                305 drivers/spi/spi-mpc52xx.c 	ms->state = mpc52xx_spi_fsmstate_transfer;
ms                314 drivers/spi/spi-mpc52xx.c static void mpc52xx_spi_fsm_process(int irq, struct mpc52xx_spi *ms)
ms                322 drivers/spi/spi-mpc52xx.c 		status = in_8(ms->regs + SPI_STATUS);
ms                323 drivers/spi/spi-mpc52xx.c 		data = in_8(ms->regs + SPI_DATA);
ms                324 drivers/spi/spi-mpc52xx.c 		rc = ms->state(irq, ms, status, data);
ms                328 drivers/spi/spi-mpc52xx.c 		schedule_work(&ms->work);
ms                336 drivers/spi/spi-mpc52xx.c 	struct mpc52xx_spi *ms = _ms;
ms                337 drivers/spi/spi-mpc52xx.c 	spin_lock(&ms->lock);
ms                338 drivers/spi/spi-mpc52xx.c 	mpc52xx_spi_fsm_process(irq, ms);
ms                339 drivers/spi/spi-mpc52xx.c 	spin_unlock(&ms->lock);
ms                348 drivers/spi/spi-mpc52xx.c 	struct mpc52xx_spi *ms = container_of(work, struct mpc52xx_spi, work);
ms                351 drivers/spi/spi-mpc52xx.c 	spin_lock_irqsave(&ms->lock, flags);
ms                352 drivers/spi/spi-mpc52xx.c 	mpc52xx_spi_fsm_process(0, ms);
ms                353 drivers/spi/spi-mpc52xx.c 	spin_unlock_irqrestore(&ms->lock, flags);
ms                362 drivers/spi/spi-mpc52xx.c 	struct mpc52xx_spi *ms = spi_master_get_devdata(spi->master);
ms                368 drivers/spi/spi-mpc52xx.c 	spin_lock_irqsave(&ms->lock, flags);
ms                369 drivers/spi/spi-mpc52xx.c 	list_add_tail(&m->queue, &ms->queue);
ms                370 drivers/spi/spi-mpc52xx.c 	spin_unlock_irqrestore(&ms->lock, flags);
ms                371 drivers/spi/spi-mpc52xx.c 	schedule_work(&ms->work);
ms                382 drivers/spi/spi-mpc52xx.c 	struct mpc52xx_spi *ms;
ms                416 drivers/spi/spi-mpc52xx.c 	master = spi_alloc_master(&op->dev, sizeof *ms);
ms                429 drivers/spi/spi-mpc52xx.c 	ms = spi_master_get_devdata(master);
ms                430 drivers/spi/spi-mpc52xx.c 	ms->master = master;
ms                431 drivers/spi/spi-mpc52xx.c 	ms->regs = regs;
ms                432 drivers/spi/spi-mpc52xx.c 	ms->irq0 = irq_of_parse_and_map(op->dev.of_node, 0);
ms                433 drivers/spi/spi-mpc52xx.c 	ms->irq1 = irq_of_parse_and_map(op->dev.of_node, 1);
ms                434 drivers/spi/spi-mpc52xx.c 	ms->state = mpc52xx_spi_fsmstate_idle;
ms                435 drivers/spi/spi-mpc52xx.c 	ms->ipb_freq = mpc5xxx_get_bus_frequency(op->dev.of_node);
ms                436 drivers/spi/spi-mpc52xx.c 	ms->gpio_cs_count = of_gpio_count(op->dev.of_node);
ms                437 drivers/spi/spi-mpc52xx.c 	if (ms->gpio_cs_count > 0) {
ms                438 drivers/spi/spi-mpc52xx.c 		master->num_chipselect = ms->gpio_cs_count;
ms                439 drivers/spi/spi-mpc52xx.c 		ms->gpio_cs = kmalloc_array(ms->gpio_cs_count,
ms                440 drivers/spi/spi-mpc52xx.c 					    sizeof(*ms->gpio_cs),
ms                442 drivers/spi/spi-mpc52xx.c 		if (!ms->gpio_cs) {
ms                447 drivers/spi/spi-mpc52xx.c 		for (i = 0; i < ms->gpio_cs_count; i++) {
ms                465 drivers/spi/spi-mpc52xx.c 			ms->gpio_cs[i] = gpio_cs;
ms                469 drivers/spi/spi-mpc52xx.c 	spin_lock_init(&ms->lock);
ms                470 drivers/spi/spi-mpc52xx.c 	INIT_LIST_HEAD(&ms->queue);
ms                471 drivers/spi/spi-mpc52xx.c 	INIT_WORK(&ms->work, mpc52xx_spi_wq);
ms                474 drivers/spi/spi-mpc52xx.c 	if (ms->irq0 && ms->irq1) {
ms                475 drivers/spi/spi-mpc52xx.c 		rc = request_irq(ms->irq0, mpc52xx_spi_irq, 0,
ms                476 drivers/spi/spi-mpc52xx.c 				  "mpc5200-spi-modf", ms);
ms                477 drivers/spi/spi-mpc52xx.c 		rc |= request_irq(ms->irq1, mpc52xx_spi_irq, 0,
ms                478 drivers/spi/spi-mpc52xx.c 				  "mpc5200-spi-spif", ms);
ms                480 drivers/spi/spi-mpc52xx.c 			free_irq(ms->irq0, ms);
ms                481 drivers/spi/spi-mpc52xx.c 			free_irq(ms->irq1, ms);
ms                482 drivers/spi/spi-mpc52xx.c 			ms->irq0 = ms->irq1 = 0;
ms                486 drivers/spi/spi-mpc52xx.c 		ms->irq0 = ms->irq1 = 0;
ms                489 drivers/spi/spi-mpc52xx.c 	if (!ms->irq0)
ms                497 drivers/spi/spi-mpc52xx.c 	dev_info(&ms->master->dev, "registered MPC5200 SPI bus\n");
ms                502 drivers/spi/spi-mpc52xx.c 	dev_err(&ms->master->dev, "initialization failed\n");
ms                505 drivers/spi/spi-mpc52xx.c 		gpio_free(ms->gpio_cs[i]);
ms                507 drivers/spi/spi-mpc52xx.c 	kfree(ms->gpio_cs);
ms                519 drivers/spi/spi-mpc52xx.c 	struct mpc52xx_spi *ms = spi_master_get_devdata(master);
ms                522 drivers/spi/spi-mpc52xx.c 	free_irq(ms->irq0, ms);
ms                523 drivers/spi/spi-mpc52xx.c 	free_irq(ms->irq1, ms);
ms                525 drivers/spi/spi-mpc52xx.c 	for (i = 0; i < ms->gpio_cs_count; i++)
ms                526 drivers/spi/spi-mpc52xx.c 		gpio_free(ms->gpio_cs[i]);
ms                528 drivers/spi/spi-mpc52xx.c 	kfree(ms->gpio_cs);
ms                530 drivers/spi/spi-mpc52xx.c 	iounmap(ms->regs);
ms                448 drivers/spi/spi-s3c64xx.c 	int ms;
ms                451 drivers/spi/spi-s3c64xx.c 	ms = xfer->len * 8 * 1000 / sdd->cur_speed;
ms                452 drivers/spi/spi-s3c64xx.c 	ms += 10; /* some tolerance */
ms                454 drivers/spi/spi-s3c64xx.c 	val = msecs_to_jiffies(ms) + 10;
ms                494 drivers/spi/spi-s3c64xx.c 	int ms;
ms                497 drivers/spi/spi-s3c64xx.c 	ms = xfer->len * 8 * 1000 / sdd->cur_speed;
ms                498 drivers/spi/spi-s3c64xx.c 	ms += 10; /* some tolerance */
ms                500 drivers/spi/spi-s3c64xx.c 	val = msecs_to_loops(ms);
ms                527 drivers/spi/spi-s3c64xx.c 						       (loops ? ms : 0));
ms               1063 drivers/spi/spi.c 	unsigned long long ms = 1;
ms               1071 drivers/spi/spi.c 		ms = 8LL * 1000LL * xfer->len;
ms               1072 drivers/spi/spi.c 		do_div(ms, xfer->speed_hz);
ms               1073 drivers/spi/spi.c 		ms += ms + 200; /* some tolerance */
ms               1075 drivers/spi/spi.c 		if (ms > UINT_MAX)
ms               1076 drivers/spi/spi.c 			ms = UINT_MAX;
ms               1078 drivers/spi/spi.c 		ms = wait_for_completion_timeout(&ctlr->xfer_completion,
ms               1079 drivers/spi/spi.c 						 msecs_to_jiffies(ms));
ms               1081 drivers/spi/spi.c 		if (ms == 0) {
ms                321 drivers/ssb/driver_chipcommon.c u32 ssb_chipco_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms)
ms                329 drivers/ssb/driver_chipcommon.c 	ticks = ssb_chipco_watchdog_timer_set(cc, cc->ticks_per_ms * ms);
ms                122 drivers/ssb/driver_extif.c u32 ssb_extif_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms)
ms                125 drivers/ssb/driver_extif.c 	u32 ticks = (SSB_EXTIF_WATCHDOG_CLK / 1000) * ms;
ms                198 drivers/ssb/ssb_private.h extern u32 ssb_chipco_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms);
ms                221 drivers/ssb/ssb_private.h extern u32 ssb_extif_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms);
ms                229 drivers/ssb/ssb_private.h 						  u32 ms)
ms                554 drivers/staging/most/configfs.c 	struct most_sound *ms = container_of(to_configfs_subsystem(group),
ms                557 drivers/staging/most/configfs.c 	list_for_each_entry(most, &ms->soundcard_list, list) {
ms                568 drivers/staging/most/configfs.c 	list_add_tail(&most->list, &ms->soundcard_list);
ms                332 drivers/staging/rtl8188eu/include/rtw_mlme.h #define rtw_set_scan_deny(adapter, ms) do {} while (0)
ms                565 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h #define set_survey_timer(mlmeext, ms) \
ms                567 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h 		  msecs_to_jiffies(ms))
ms                569 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h #define set_link_timer(mlmeext, ms) \
ms                571 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h 		  msecs_to_jiffies(ms))
ms                211 drivers/staging/rtl8188eu/include/rtw_pwrctrl.h #define _rtw_set_pwr_state_check_timer(pwrctrlpriv, ms) \
ms                213 drivers/staging/rtl8188eu/include/rtw_pwrctrl.h 		  jiffies + msecs_to_jiffies(ms))
ms               1934 drivers/staging/rtl8723bs/core/rtw_mlme.c void rtw_set_scan_deny(struct adapter *adapter, u32 ms)
ms               1940 drivers/staging/rtl8723bs/core/rtw_mlme.c 	_set_timer(&mlmepriv->set_scan_deny_timer, ms);
ms               1188 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c inline void rtw_set_ips_deny(struct adapter *padapter, u32 ms)
ms               1191 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c 	pwrpriv->ips_deny_time = jiffies + msecs_to_jiffies(ms);
ms                459 drivers/staging/rtl8723bs/include/rtw_mlme.h #define rtw_mlme_set_auto_scan_int(adapter, ms) \
ms                461 drivers/staging/rtl8723bs/include/rtw_mlme.h 		adapter->mlmepriv.auto_scan_int_ms = ms; \
ms                590 drivers/staging/rtl8723bs/include/rtw_mlme.h void rtw_set_scan_deny(struct adapter *adapter, u32 ms);
ms                719 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_survey_timer(mlmeext, ms) \
ms                722 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h 		_set_timer(&(mlmeext)->survey_timer, (ms)); \
ms                725 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_link_timer(mlmeext, ms) \
ms                728 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h 		_set_timer(&(mlmeext)->link_timer, (ms)); \
ms                730 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h #define set_sa_query_timer(mlmeext, ms) \
ms                732 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h 		DBG_871X("%s set_sa_query_timer(%p, %d)\n", __func__, (mlmeext), (ms)); \
ms                733 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h 		_set_timer(&(mlmeext)->sa_query_timer, (ms)); \
ms                315 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h #define _rtw_set_pwr_state_check_timer(pwrctl, ms) \
ms                318 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h 		_set_timer(&(pwrctl)->pwr_state_check_timer, (ms)); \
ms                352 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h void rtw_set_ips_deny(struct adapter *padapter, u32 ms);
ms                302 drivers/thermal/intel/x86_pkg_temp_thermal.c 	unsigned long ms = msecs_to_jiffies(notify_delay_ms);
ms                304 drivers/thermal/intel/x86_pkg_temp_thermal.c 	schedule_delayed_work_on(cpu, work, ms);
ms                248 drivers/tty/hvc/hvsi_lib.c static void maybe_msleep(unsigned long ms)
ms                252 drivers/tty/hvc/hvsi_lib.c 		mdelay(ms);
ms                254 drivers/tty/hvc/hvsi_lib.c 		msleep(ms);
ms               1525 drivers/tty/mxser.c 		struct mxser_mstatus ms, __user *msu = argp;
ms               1530 drivers/tty/mxser.c 				memset(&ms, 0, sizeof(ms));
ms               1539 drivers/tty/mxser.c 					ms.cflag = ip->normal_termios.c_cflag;
ms               1541 drivers/tty/mxser.c 					ms.cflag = tty->termios.c_cflag;
ms               1547 drivers/tty/mxser.c 					ms.dcd = 1;
ms               1549 drivers/tty/mxser.c 					ms.dsr = 1;
ms               1551 drivers/tty/mxser.c 					ms.cts = 1;
ms               1554 drivers/tty/mxser.c 				if (copy_to_user(msu, &ms, sizeof(ms)))
ms               1401 drivers/tty/serial/sunsu.c 		const char *ms = of_get_property(ap, "mouse", NULL);
ms               1419 drivers/tty/serial/sunsu.c 		if (ms) {
ms               1420 drivers/tty/serial/sunsu.c 			match = of_find_node_by_path(ms);
ms                147 drivers/usb/class/cdc-acm.c #define acm_send_break(acm, ms) \
ms                148 drivers/usb/class/cdc-acm.c 	acm_ctrl_msg(acm, USB_CDC_REQ_SEND_BREAK, ms, NULL, 0)
ms                398 drivers/usb/dwc2/core.c 		s64 ms;
ms                407 drivers/usb/dwc2/core.c 		ms = ktime_to_ms(ktime_sub(end, start));
ms                409 drivers/usb/dwc2/core.c 		if (ms >= (s64)timeout) {
ms                185 drivers/usb/early/ehci-dbgp.c static inline void dbgp_mdelay(int ms)
ms                189 drivers/usb/early/ehci-dbgp.c 	while (ms--) {
ms                581 drivers/usb/storage/scsiglue.c 	unsigned short ms;
ms                583 drivers/usb/storage/scsiglue.c 	if (sscanf(buf, "%hu", &ms) > 0) {
ms                584 drivers/usb/storage/scsiglue.c 		blk_queue_max_hw_sectors(sdev->request_queue, ms);
ms                 42 drivers/video/fbdev/amba-clcd.c static inline void clcdfb_sleep(unsigned int ms)
ms                 45 drivers/video/fbdev/amba-clcd.c 		mdelay(ms);
ms                 47 drivers/video/fbdev/amba-clcd.c 		msleep(ms);
ms                284 drivers/video/fbdev/aty/radeon_base.c void _radeon_msleep(struct radeonfb_info *rinfo, unsigned long ms)
ms                287 drivers/video/fbdev/aty/radeon_base.c 		mdelay(ms);
ms                289 drivers/video/fbdev/aty/radeon_base.c 		msleep(ms);
ms                374 drivers/video/fbdev/aty/radeonfb.h void _radeon_msleep(struct radeonfb_info *rinfo, unsigned long ms);
ms                477 drivers/video/fbdev/aty/radeonfb.h #define radeon_msleep(ms)		_radeon_msleep(rinfo,ms)
ms                511 drivers/w1/masters/ds2490.c 	int ms = del<<4;
ms                520 drivers/w1/masters/ds2490.c 	if (delay == 0 || ms == dev->spu_sleep)
ms                527 drivers/w1/masters/ds2490.c 	dev->spu_sleep = ms;
ms                 87 fs/dlm/lock.c  				    struct dlm_message *ms);
ms                 88 fs/dlm/lock.c  static int receive_extralen(struct dlm_message *ms);
ms               1509 fs/dlm/lock.c  				struct dlm_message *ms)
ms               1548 fs/dlm/lock.c  	    is_overlap_cancel(lkb) && ms && !ms->m_result) {
ms               1566 fs/dlm/lock.c  		  lkb->lkb_id, ms ? ms->m_header.h_nodeid : 0, lkb->lkb_remid,
ms               1607 fs/dlm/lock.c  static int remove_from_waiters_ms(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               1612 fs/dlm/lock.c  	if (ms->m_flags != DLM_IFL_STUB_MS)
ms               1614 fs/dlm/lock.c  	error = _remove_from_waiters(lkb, ms->m_type, ms);
ms               1615 fs/dlm/lock.c  	if (ms->m_flags != DLM_IFL_STUB_MS)
ms               2026 fs/dlm/lock.c  			    struct dlm_message *ms)
ms               2038 fs/dlm/lock.c  		int len = receive_extralen(ms);
ms               2041 fs/dlm/lock.c  		memcpy(lkb->lkb_lvbptr, ms->m_extra, len);
ms               2042 fs/dlm/lock.c  		lkb->lkb_lvbseq = ms->m_lvbseq;
ms               2133 fs/dlm/lock.c  			  struct dlm_message *ms)
ms               2135 fs/dlm/lock.c  	set_lvb_lock_pc(r, lkb, ms);
ms               2171 fs/dlm/lock.c  static void munge_altmode(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               2173 fs/dlm/lock.c  	if (ms->m_type != DLM_MSG_REQUEST_REPLY &&
ms               2174 fs/dlm/lock.c  	    ms->m_type != DLM_MSG_GRANT) {
ms               2176 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_type);
ms               3532 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3546 fs/dlm/lock.c  	ms = (struct dlm_message *) mb;
ms               3548 fs/dlm/lock.c  	ms->m_header.h_version = (DLM_HEADER_MAJOR | DLM_HEADER_MINOR);
ms               3549 fs/dlm/lock.c  	ms->m_header.h_lockspace = ls->ls_global_id;
ms               3550 fs/dlm/lock.c  	ms->m_header.h_nodeid = dlm_our_nodeid();
ms               3551 fs/dlm/lock.c  	ms->m_header.h_length = mb_len;
ms               3552 fs/dlm/lock.c  	ms->m_header.h_cmd = DLM_MSG;
ms               3554 fs/dlm/lock.c  	ms->m_type = mstype;
ms               3557 fs/dlm/lock.c  	*ms_ret = ms;
ms               3591 fs/dlm/lock.c  static int send_message(struct dlm_mhandle *mh, struct dlm_message *ms)
ms               3593 fs/dlm/lock.c  	dlm_message_out(ms);
ms               3599 fs/dlm/lock.c  		      struct dlm_message *ms)
ms               3601 fs/dlm/lock.c  	ms->m_nodeid   = lkb->lkb_nodeid;
ms               3602 fs/dlm/lock.c  	ms->m_pid      = lkb->lkb_ownpid;
ms               3603 fs/dlm/lock.c  	ms->m_lkid     = lkb->lkb_id;
ms               3604 fs/dlm/lock.c  	ms->m_remid    = lkb->lkb_remid;
ms               3605 fs/dlm/lock.c  	ms->m_exflags  = lkb->lkb_exflags;
ms               3606 fs/dlm/lock.c  	ms->m_sbflags  = lkb->lkb_sbflags;
ms               3607 fs/dlm/lock.c  	ms->m_flags    = lkb->lkb_flags;
ms               3608 fs/dlm/lock.c  	ms->m_lvbseq   = lkb->lkb_lvbseq;
ms               3609 fs/dlm/lock.c  	ms->m_status   = lkb->lkb_status;
ms               3610 fs/dlm/lock.c  	ms->m_grmode   = lkb->lkb_grmode;
ms               3611 fs/dlm/lock.c  	ms->m_rqmode   = lkb->lkb_rqmode;
ms               3612 fs/dlm/lock.c  	ms->m_hash     = r->res_hash;
ms               3618 fs/dlm/lock.c  		ms->m_asts |= DLM_CB_BAST;
ms               3620 fs/dlm/lock.c  		ms->m_asts |= DLM_CB_CAST;
ms               3625 fs/dlm/lock.c  	switch (ms->m_type) {
ms               3628 fs/dlm/lock.c  		memcpy(ms->m_extra, r->res_name, r->res_length);
ms               3637 fs/dlm/lock.c  		memcpy(ms->m_extra, lkb->lkb_lvbptr, r->res_ls->ls_lvblen);
ms               3644 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3654 fs/dlm/lock.c  	error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh);
ms               3658 fs/dlm/lock.c  	send_args(r, lkb, ms);
ms               3660 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3709 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3715 fs/dlm/lock.c  	error = create_message(r, lkb, to_nodeid, DLM_MSG_GRANT, &ms, &mh);
ms               3719 fs/dlm/lock.c  	send_args(r, lkb, ms);
ms               3721 fs/dlm/lock.c  	ms->m_result = 0;
ms               3723 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3730 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3736 fs/dlm/lock.c  	error = create_message(r, NULL, to_nodeid, DLM_MSG_BAST, &ms, &mh);
ms               3740 fs/dlm/lock.c  	send_args(r, lkb, ms);
ms               3742 fs/dlm/lock.c  	ms->m_bastmode = mode;
ms               3744 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3751 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3761 fs/dlm/lock.c  	error = create_message(r, NULL, to_nodeid, DLM_MSG_LOOKUP, &ms, &mh);
ms               3765 fs/dlm/lock.c  	send_args(r, lkb, ms);
ms               3767 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3779 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3785 fs/dlm/lock.c  	error = create_message(r, NULL, to_nodeid, DLM_MSG_REMOVE, &ms, &mh);
ms               3789 fs/dlm/lock.c  	memcpy(ms->m_extra, r->res_name, r->res_length);
ms               3790 fs/dlm/lock.c  	ms->m_hash = r->res_hash;
ms               3792 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3800 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3806 fs/dlm/lock.c  	error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh);
ms               3810 fs/dlm/lock.c  	send_args(r, lkb, ms);
ms               3812 fs/dlm/lock.c  	ms->m_result = rv;
ms               3814 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3843 fs/dlm/lock.c  	struct dlm_message *ms;
ms               3847 fs/dlm/lock.c  	error = create_message(r, NULL, nodeid, DLM_MSG_LOOKUP_REPLY, &ms, &mh);
ms               3851 fs/dlm/lock.c  	ms->m_lkid = ms_in->m_lkid;
ms               3852 fs/dlm/lock.c  	ms->m_result = rv;
ms               3853 fs/dlm/lock.c  	ms->m_nodeid = ret_nodeid;
ms               3855 fs/dlm/lock.c  	error = send_message(mh, ms);
ms               3864 fs/dlm/lock.c  static void receive_flags(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               3866 fs/dlm/lock.c  	lkb->lkb_exflags = ms->m_exflags;
ms               3867 fs/dlm/lock.c  	lkb->lkb_sbflags = ms->m_sbflags;
ms               3869 fs/dlm/lock.c  		         (ms->m_flags & 0x0000FFFF);
ms               3872 fs/dlm/lock.c  static void receive_flags_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               3874 fs/dlm/lock.c  	if (ms->m_flags == DLM_IFL_STUB_MS)
ms               3877 fs/dlm/lock.c  	lkb->lkb_sbflags = ms->m_sbflags;
ms               3879 fs/dlm/lock.c  		         (ms->m_flags & 0x0000FFFF);
ms               3882 fs/dlm/lock.c  static int receive_extralen(struct dlm_message *ms)
ms               3884 fs/dlm/lock.c  	return (ms->m_header.h_length - sizeof(struct dlm_message));
ms               3888 fs/dlm/lock.c  		       struct dlm_message *ms)
ms               3897 fs/dlm/lock.c  		len = receive_extralen(ms);
ms               3900 fs/dlm/lock.c  		memcpy(lkb->lkb_lvbptr, ms->m_extra, len);
ms               3916 fs/dlm/lock.c  				struct dlm_message *ms)
ms               3918 fs/dlm/lock.c  	lkb->lkb_nodeid = ms->m_header.h_nodeid;
ms               3919 fs/dlm/lock.c  	lkb->lkb_ownpid = ms->m_pid;
ms               3920 fs/dlm/lock.c  	lkb->lkb_remid = ms->m_lkid;
ms               3922 fs/dlm/lock.c  	lkb->lkb_rqmode = ms->m_rqmode;
ms               3924 fs/dlm/lock.c  	lkb->lkb_bastfn = (ms->m_asts & DLM_CB_BAST) ? &fake_bastfn : NULL;
ms               3925 fs/dlm/lock.c  	lkb->lkb_astfn = (ms->m_asts & DLM_CB_CAST) ? &fake_astfn : NULL;
ms               3938 fs/dlm/lock.c  				struct dlm_message *ms)
ms               3943 fs/dlm/lock.c  	if (receive_lvb(ls, lkb, ms))
ms               3946 fs/dlm/lock.c  	lkb->lkb_rqmode = ms->m_rqmode;
ms               3947 fs/dlm/lock.c  	lkb->lkb_lvbseq = ms->m_lvbseq;
ms               3953 fs/dlm/lock.c  			       struct dlm_message *ms)
ms               3955 fs/dlm/lock.c  	if (receive_lvb(ls, lkb, ms))
ms               3963 fs/dlm/lock.c  static void setup_stub_lkb(struct dlm_ls *ls, struct dlm_message *ms)
ms               3966 fs/dlm/lock.c  	lkb->lkb_nodeid = ms->m_header.h_nodeid;
ms               3967 fs/dlm/lock.c  	lkb->lkb_remid = ms->m_lkid;
ms               3973 fs/dlm/lock.c  static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               3975 fs/dlm/lock.c  	int from = ms->m_header.h_nodeid;
ms               3978 fs/dlm/lock.c  	switch (ms->m_type) {
ms               4009 fs/dlm/lock.c  			  ms->m_type, from, lkb->lkb_id, lkb->lkb_remid,
ms               4017 fs/dlm/lock.c  	struct dlm_message *ms;
ms               4057 fs/dlm/lock.c  			     dir_nodeid, DLM_MSG_REMOVE, &ms, &mh);
ms               4061 fs/dlm/lock.c  	memcpy(ms->m_extra, name, len);
ms               4062 fs/dlm/lock.c  	ms->m_hash = hash;
ms               4064 fs/dlm/lock.c  	send_message(mh, ms);
ms               4072 fs/dlm/lock.c  static int receive_request(struct dlm_ls *ls, struct dlm_message *ms)
ms               4079 fs/dlm/lock.c  	from_nodeid = ms->m_header.h_nodeid;
ms               4085 fs/dlm/lock.c  	receive_flags(lkb, ms);
ms               4087 fs/dlm/lock.c  	error = receive_request_args(ls, lkb, ms);
ms               4099 fs/dlm/lock.c  	namelen = receive_extralen(ms);
ms               4101 fs/dlm/lock.c  	error = find_rsb(ls, ms->m_extra, namelen, from_nodeid,
ms               4152 fs/dlm/lock.c  			  ms->m_lkid, from_nodeid, error);
ms               4156 fs/dlm/lock.c  		send_repeat_remove(ls, ms->m_extra, namelen);
ms               4160 fs/dlm/lock.c  	setup_stub_lkb(ls, ms);
ms               4165 fs/dlm/lock.c  static int receive_convert(struct dlm_ls *ls, struct dlm_message *ms)
ms               4171 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4175 fs/dlm/lock.c  	if (lkb->lkb_remid != ms->m_lkid) {
ms               4179 fs/dlm/lock.c  			  ms->m_header.h_nodeid, ms->m_lkid);
ms               4190 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4194 fs/dlm/lock.c  	receive_flags(lkb, ms);
ms               4196 fs/dlm/lock.c  	error = receive_convert_args(ls, lkb, ms);
ms               4215 fs/dlm/lock.c  	setup_stub_lkb(ls, ms);
ms               4220 fs/dlm/lock.c  static int receive_unlock(struct dlm_ls *ls, struct dlm_message *ms)
ms               4226 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4230 fs/dlm/lock.c  	if (lkb->lkb_remid != ms->m_lkid) {
ms               4233 fs/dlm/lock.c  			  ms->m_header.h_nodeid, ms->m_lkid);
ms               4244 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4248 fs/dlm/lock.c  	receive_flags(lkb, ms);
ms               4250 fs/dlm/lock.c  	error = receive_unlock_args(ls, lkb, ms);
ms               4266 fs/dlm/lock.c  	setup_stub_lkb(ls, ms);
ms               4271 fs/dlm/lock.c  static int receive_cancel(struct dlm_ls *ls, struct dlm_message *ms)
ms               4277 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4281 fs/dlm/lock.c  	receive_flags(lkb, ms);
ms               4288 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4302 fs/dlm/lock.c  	setup_stub_lkb(ls, ms);
ms               4307 fs/dlm/lock.c  static int receive_grant(struct dlm_ls *ls, struct dlm_message *ms)
ms               4313 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4322 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4326 fs/dlm/lock.c  	receive_flags_reply(lkb, ms);
ms               4328 fs/dlm/lock.c  		munge_altmode(lkb, ms);
ms               4329 fs/dlm/lock.c  	grant_lock_pc(r, lkb, ms);
ms               4338 fs/dlm/lock.c  static int receive_bast(struct dlm_ls *ls, struct dlm_message *ms)
ms               4344 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4353 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4357 fs/dlm/lock.c  	queue_bast(r, lkb, ms->m_bastmode);
ms               4358 fs/dlm/lock.c  	lkb->lkb_highbast = ms->m_bastmode;
ms               4366 fs/dlm/lock.c  static void receive_lookup(struct dlm_ls *ls, struct dlm_message *ms)
ms               4370 fs/dlm/lock.c  	from_nodeid = ms->m_header.h_nodeid;
ms               4373 fs/dlm/lock.c  	len = receive_extralen(ms);
ms               4375 fs/dlm/lock.c  	error = dlm_master_lookup(ls, from_nodeid, ms->m_extra, len, 0,
ms               4380 fs/dlm/lock.c  		receive_request(ls, ms);
ms               4383 fs/dlm/lock.c  	send_lookup_reply(ls, ms, ret_nodeid, error);
ms               4386 fs/dlm/lock.c  static void receive_remove(struct dlm_ls *ls, struct dlm_message *ms)
ms               4393 fs/dlm/lock.c  	from_nodeid = ms->m_header.h_nodeid;
ms               4395 fs/dlm/lock.c  	len = receive_extralen(ms);
ms               4403 fs/dlm/lock.c  	dir_nodeid = dlm_hash2nodeid(ls, ms->m_hash);
ms               4420 fs/dlm/lock.c  	memcpy(name, ms->m_extra, len);
ms               4474 fs/dlm/lock.c  static void receive_purge(struct dlm_ls *ls, struct dlm_message *ms)
ms               4476 fs/dlm/lock.c  	do_purge(ls, ms->m_nodeid, ms->m_pid);
ms               4479 fs/dlm/lock.c  static int receive_request_reply(struct dlm_ls *ls, struct dlm_message *ms)
ms               4484 fs/dlm/lock.c  	int from_nodeid = ms->m_header.h_nodeid;
ms               4486 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4494 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4502 fs/dlm/lock.c  			  lkb->lkb_id, from_nodeid, ms->m_lkid, ms->m_result);
ms               4516 fs/dlm/lock.c  	result = ms->m_result;
ms               4529 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4530 fs/dlm/lock.c  		lkb->lkb_remid = ms->m_lkid;
ms               4532 fs/dlm/lock.c  			munge_altmode(lkb, ms);
ms               4537 fs/dlm/lock.c  			grant_lock_pc(r, lkb, ms);
ms               4600 fs/dlm/lock.c  				    struct dlm_message *ms)
ms               4603 fs/dlm/lock.c  	switch (ms->m_result) {
ms               4610 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4617 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4627 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4630 fs/dlm/lock.c  		grant_lock_pc(r, lkb, ms);
ms               4636 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_header.h_nodeid, ms->m_lkid,
ms               4637 fs/dlm/lock.c  			  ms->m_result);
ms               4643 fs/dlm/lock.c  static void _receive_convert_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               4651 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4656 fs/dlm/lock.c  	error = remove_from_waiters_ms(lkb, ms);
ms               4660 fs/dlm/lock.c  	__receive_convert_reply(r, lkb, ms);
ms               4666 fs/dlm/lock.c  static int receive_convert_reply(struct dlm_ls *ls, struct dlm_message *ms)
ms               4671 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4675 fs/dlm/lock.c  	_receive_convert_reply(lkb, ms);
ms               4680 fs/dlm/lock.c  static void _receive_unlock_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               4688 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4693 fs/dlm/lock.c  	error = remove_from_waiters_ms(lkb, ms);
ms               4699 fs/dlm/lock.c  	switch (ms->m_result) {
ms               4701 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4709 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_result);
ms               4716 fs/dlm/lock.c  static int receive_unlock_reply(struct dlm_ls *ls, struct dlm_message *ms)
ms               4721 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4725 fs/dlm/lock.c  	_receive_unlock_reply(lkb, ms);
ms               4730 fs/dlm/lock.c  static void _receive_cancel_reply(struct dlm_lkb *lkb, struct dlm_message *ms)
ms               4738 fs/dlm/lock.c  	error = validate_message(lkb, ms);
ms               4743 fs/dlm/lock.c  	error = remove_from_waiters_ms(lkb, ms);
ms               4749 fs/dlm/lock.c  	switch (ms->m_result) {
ms               4751 fs/dlm/lock.c  		receive_flags_reply(lkb, ms);
ms               4759 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_result);
ms               4766 fs/dlm/lock.c  static int receive_cancel_reply(struct dlm_ls *ls, struct dlm_message *ms)
ms               4771 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_remid, &lkb);
ms               4775 fs/dlm/lock.c  	_receive_cancel_reply(lkb, ms);
ms               4780 fs/dlm/lock.c  static void receive_lookup_reply(struct dlm_ls *ls, struct dlm_message *ms)
ms               4787 fs/dlm/lock.c  	error = find_lkb(ls, ms->m_lkid, &lkb);
ms               4789 fs/dlm/lock.c  		log_error(ls, "receive_lookup_reply no lkid %x", ms->m_lkid);
ms               4804 fs/dlm/lock.c  	ret_nodeid = ms->m_nodeid;
ms               4816 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_header.h_nodeid, ret_nodeid,
ms               4829 fs/dlm/lock.c  			  lkb->lkb_id, ms->m_header.h_nodeid);
ms               4858 fs/dlm/lock.c  static void _receive_message(struct dlm_ls *ls, struct dlm_message *ms,
ms               4863 fs/dlm/lock.c  	if (!dlm_is_member(ls, ms->m_header.h_nodeid)) {
ms               4865 fs/dlm/lock.c  			  ms->m_type, ms->m_header.h_nodeid, ms->m_lkid,
ms               4866 fs/dlm/lock.c  			  ms->m_remid, ms->m_result);
ms               4870 fs/dlm/lock.c  	switch (ms->m_type) {
ms               4875 fs/dlm/lock.c  		error = receive_request(ls, ms);
ms               4879 fs/dlm/lock.c  		error = receive_convert(ls, ms);
ms               4883 fs/dlm/lock.c  		error = receive_unlock(ls, ms);
ms               4888 fs/dlm/lock.c  		error = receive_cancel(ls, ms);
ms               4894 fs/dlm/lock.c  		error = receive_request_reply(ls, ms);
ms               4898 fs/dlm/lock.c  		error = receive_convert_reply(ls, ms);
ms               4902 fs/dlm/lock.c  		error = receive_unlock_reply(ls, ms);
ms               4906 fs/dlm/lock.c  		error = receive_cancel_reply(ls, ms);
ms               4913 fs/dlm/lock.c  		error = receive_grant(ls, ms);
ms               4918 fs/dlm/lock.c  		error = receive_bast(ls, ms);
ms               4924 fs/dlm/lock.c  		receive_lookup(ls, ms);
ms               4928 fs/dlm/lock.c  		receive_remove(ls, ms);
ms               4934 fs/dlm/lock.c  		receive_lookup_reply(ls, ms);
ms               4940 fs/dlm/lock.c  		receive_purge(ls, ms);
ms               4944 fs/dlm/lock.c  		log_error(ls, "unknown message type %d", ms->m_type);
ms               4960 fs/dlm/lock.c  			  ms->m_type, ms->m_remid, ms->m_header.h_nodeid,
ms               4961 fs/dlm/lock.c  			  ms->m_lkid, saved_seq);
ms               4964 fs/dlm/lock.c  			  ms->m_type, ms->m_remid, ms->m_header.h_nodeid,
ms               4965 fs/dlm/lock.c  			  ms->m_lkid, saved_seq);
ms               4967 fs/dlm/lock.c  		if (ms->m_type == DLM_MSG_CONVERT)
ms               4968 fs/dlm/lock.c  			dlm_dump_rsb_hash(ls, ms->m_hash);
ms               4974 fs/dlm/lock.c  			  ms->m_type, ms->m_header.h_nodeid,
ms               4975 fs/dlm/lock.c  			  ms->m_lkid, ms->m_remid, saved_seq);
ms               4987 fs/dlm/lock.c  static void dlm_receive_message(struct dlm_ls *ls, struct dlm_message *ms,
ms               4996 fs/dlm/lock.c  				  ms->m_type, nodeid);
ms               5000 fs/dlm/lock.c  		dlm_add_requestqueue(ls, nodeid, ms);
ms               5003 fs/dlm/lock.c  		_receive_message(ls, ms, 0);
ms               5010 fs/dlm/lock.c  void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms,
ms               5013 fs/dlm/lock.c  	_receive_message(ls, ms, saved_seq);
ms               6274 fs/dlm/lock.c  	struct dlm_message *ms;
ms               6279 fs/dlm/lock.c  				DLM_MSG_PURGE, &ms, &mh);
ms               6282 fs/dlm/lock.c  	ms->m_nodeid = nodeid;
ms               6283 fs/dlm/lock.c  	ms->m_pid = pid;
ms               6285 fs/dlm/lock.c  	return send_message(mh, ms);
ms                 17 fs/dlm/lock.h  void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms,
ms                 32 fs/dlm/requestqueue.c void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms)
ms                 35 fs/dlm/requestqueue.c 	int length = ms->m_header.h_length - sizeof(struct dlm_message);
ms                 45 fs/dlm/requestqueue.c 	memcpy(&e->request, ms, ms->m_header.h_length);
ms                 66 fs/dlm/requestqueue.c 	struct dlm_message *ms;
ms                 80 fs/dlm/requestqueue.c 		ms = &e->request;
ms                 84 fs/dlm/requestqueue.c 			  ms->m_type, ms->m_header.h_nodeid,
ms                 85 fs/dlm/requestqueue.c 			  ms->m_lkid, ms->m_remid, ms->m_result,
ms                128 fs/dlm/requestqueue.c static int purge_request(struct dlm_ls *ls, struct dlm_message *ms, int nodeid)
ms                130 fs/dlm/requestqueue.c 	uint32_t type = ms->m_type;
ms                155 fs/dlm/requestqueue.c 	struct dlm_message *ms;
ms                160 fs/dlm/requestqueue.c 		ms =  &e->request;
ms                162 fs/dlm/requestqueue.c 		if (purge_request(ls, ms, e->nodeid)) {
ms                 14 fs/dlm/requestqueue.h void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms);
ms                 84 fs/dlm/util.c  void dlm_message_out(struct dlm_message *ms)
ms                 86 fs/dlm/util.c  	header_out(&ms->m_header);
ms                 88 fs/dlm/util.c  	ms->m_type		= cpu_to_le32(ms->m_type);
ms                 89 fs/dlm/util.c  	ms->m_nodeid		= cpu_to_le32(ms->m_nodeid);
ms                 90 fs/dlm/util.c  	ms->m_pid		= cpu_to_le32(ms->m_pid);
ms                 91 fs/dlm/util.c  	ms->m_lkid		= cpu_to_le32(ms->m_lkid);
ms                 92 fs/dlm/util.c  	ms->m_remid		= cpu_to_le32(ms->m_remid);
ms                 93 fs/dlm/util.c  	ms->m_parent_lkid	= cpu_to_le32(ms->m_parent_lkid);
ms                 94 fs/dlm/util.c  	ms->m_parent_remid	= cpu_to_le32(ms->m_parent_remid);
ms                 95 fs/dlm/util.c  	ms->m_exflags		= cpu_to_le32(ms->m_exflags);
ms                 96 fs/dlm/util.c  	ms->m_sbflags		= cpu_to_le32(ms->m_sbflags);
ms                 97 fs/dlm/util.c  	ms->m_flags		= cpu_to_le32(ms->m_flags);
ms                 98 fs/dlm/util.c  	ms->m_lvbseq		= cpu_to_le32(ms->m_lvbseq);
ms                 99 fs/dlm/util.c  	ms->m_hash		= cpu_to_le32(ms->m_hash);
ms                100 fs/dlm/util.c  	ms->m_status		= cpu_to_le32(ms->m_status);
ms                101 fs/dlm/util.c  	ms->m_grmode		= cpu_to_le32(ms->m_grmode);
ms                102 fs/dlm/util.c  	ms->m_rqmode		= cpu_to_le32(ms->m_rqmode);
ms                103 fs/dlm/util.c  	ms->m_bastmode		= cpu_to_le32(ms->m_bastmode);
ms                104 fs/dlm/util.c  	ms->m_asts		= cpu_to_le32(ms->m_asts);
ms                105 fs/dlm/util.c  	ms->m_result		= cpu_to_le32(to_dlm_errno(ms->m_result));
ms                108 fs/dlm/util.c  void dlm_message_in(struct dlm_message *ms)
ms                110 fs/dlm/util.c  	header_in(&ms->m_header);
ms                112 fs/dlm/util.c  	ms->m_type		= le32_to_cpu(ms->m_type);
ms                113 fs/dlm/util.c  	ms->m_nodeid		= le32_to_cpu(ms->m_nodeid);
ms                114 fs/dlm/util.c  	ms->m_pid		= le32_to_cpu(ms->m_pid);
ms                115 fs/dlm/util.c  	ms->m_lkid		= le32_to_cpu(ms->m_lkid);
ms                116 fs/dlm/util.c  	ms->m_remid		= le32_to_cpu(ms->m_remid);
ms                117 fs/dlm/util.c  	ms->m_parent_lkid	= le32_to_cpu(ms->m_parent_lkid);
ms                118 fs/dlm/util.c  	ms->m_parent_remid	= le32_to_cpu(ms->m_parent_remid);
ms                119 fs/dlm/util.c  	ms->m_exflags		= le32_to_cpu(ms->m_exflags);
ms                120 fs/dlm/util.c  	ms->m_sbflags		= le32_to_cpu(ms->m_sbflags);
ms                121 fs/dlm/util.c  	ms->m_flags		= le32_to_cpu(ms->m_flags);
ms                122 fs/dlm/util.c  	ms->m_lvbseq		= le32_to_cpu(ms->m_lvbseq);
ms                123 fs/dlm/util.c  	ms->m_hash		= le32_to_cpu(ms->m_hash);
ms                124 fs/dlm/util.c  	ms->m_status		= le32_to_cpu(ms->m_status);
ms                125 fs/dlm/util.c  	ms->m_grmode		= le32_to_cpu(ms->m_grmode);
ms                126 fs/dlm/util.c  	ms->m_rqmode		= le32_to_cpu(ms->m_rqmode);
ms                127 fs/dlm/util.c  	ms->m_bastmode		= le32_to_cpu(ms->m_bastmode);
ms                128 fs/dlm/util.c  	ms->m_asts		= le32_to_cpu(ms->m_asts);
ms                129 fs/dlm/util.c  	ms->m_result		= from_dlm_errno(le32_to_cpu(ms->m_result));
ms                 14 fs/dlm/util.h  void dlm_message_out(struct dlm_message *ms);
ms                 15 fs/dlm/util.h  void dlm_message_in(struct dlm_message *ms);
ms               1797 fs/eventpoll.c static inline struct timespec64 ep_set_mstimeout(long ms)
ms               1800 fs/eventpoll.c 		.tv_sec = ms / MSEC_PER_SEC,
ms               1801 fs/eventpoll.c 		.tv_nsec = NSEC_PER_MSEC * (ms % MSEC_PER_SEC),
ms                176 fs/isofs/inode.c isofs_hashi_common(const struct dentry *dentry, struct qstr *qstr, int ms)
ms                185 fs/isofs/inode.c 	if (ms) {
ms                205 fs/isofs/inode.c 		const struct qstr *name, int ms, int ci)
ms                212 fs/isofs/inode.c 	if (ms) {
ms                248 fs/isofs/inode.c isofs_hash_common(const struct dentry *dentry, struct qstr *qstr, int ms)
ms                255 fs/isofs/inode.c 	if (ms) {
ms                119 fs/minix/inode.c 	struct minix_super_block * ms;
ms                122 fs/minix/inode.c 	ms = sbi->s_ms;
ms                126 fs/minix/inode.c 		if (ms->s_state & MINIX_VALID_FS ||
ms                131 fs/minix/inode.c 			ms->s_state = sbi->s_mount_state;
ms                136 fs/minix/inode.c 			sbi->s_mount_state = ms->s_state;
ms                137 fs/minix/inode.c 			ms->s_state &= ~MINIX_VALID_FS;
ms                157 fs/minix/inode.c 	struct minix_super_block *ms;
ms                178 fs/minix/inode.c 	ms = (struct minix_super_block *) bh->b_data;
ms                179 fs/minix/inode.c 	sbi->s_ms = ms;
ms                181 fs/minix/inode.c 	sbi->s_mount_state = ms->s_state;
ms                182 fs/minix/inode.c 	sbi->s_ninodes = ms->s_ninodes;
ms                183 fs/minix/inode.c 	sbi->s_nzones = ms->s_nzones;
ms                184 fs/minix/inode.c 	sbi->s_imap_blocks = ms->s_imap_blocks;
ms                185 fs/minix/inode.c 	sbi->s_zmap_blocks = ms->s_zmap_blocks;
ms                186 fs/minix/inode.c 	sbi->s_firstdatazone = ms->s_firstdatazone;
ms                187 fs/minix/inode.c 	sbi->s_log_zone_size = ms->s_log_zone_size;
ms                188 fs/minix/inode.c 	sbi->s_max_size = ms->s_max_size;
ms                189 fs/minix/inode.c 	s->s_magic = ms->s_magic;
ms                202 fs/minix/inode.c 		sbi->s_nzones = ms->s_zones;
ms                208 fs/minix/inode.c 		sbi->s_nzones = ms->s_zones;
ms                295 fs/minix/inode.c 			ms->s_state &= ~MINIX_VALID_FS;
ms                 37 include/asm-generic/mmiowb.h 	struct mmiowb_state *ms = __mmiowb_state();
ms                 38 include/asm-generic/mmiowb.h 	ms->mmiowb_pending = ms->nesting_count;
ms                 43 include/asm-generic/mmiowb.h 	struct mmiowb_state *ms = __mmiowb_state();
ms                 44 include/asm-generic/mmiowb.h 	ms->nesting_count++;
ms                 49 include/asm-generic/mmiowb.h 	struct mmiowb_state *ms = __mmiowb_state();
ms                 51 include/asm-generic/mmiowb.h 	if (unlikely(ms->mmiowb_pending)) {
ms                 52 include/asm-generic/mmiowb.h 		ms->mmiowb_pending = 0;
ms                 56 include/asm-generic/mmiowb.h 	ms->nesting_count--;
ms                148 include/linux/ceph/libceph.h #define from_msgr(ms)	container_of(ms, struct ceph_client, msgr)
ms                270 include/linux/ktime.h static inline ktime_t ms_to_ktime(u64 ms)
ms                272 include/linux/ktime.h 	return ms * NSEC_PER_MSEC;
ms                354 include/linux/memory_hotplug.h extern void sparse_remove_section(struct mem_section *ms,
ms               1230 include/linux/mmzone.h static inline unsigned long *section_to_usemap(struct mem_section *ms)
ms               1232 include/linux/mmzone.h 	return ms->usage->pageblock_flags;
ms               1245 include/linux/mmzone.h extern unsigned long __section_nr(struct mem_section *ms);
ms               1331 include/linux/mmzone.h static inline int pfn_section_valid(struct mem_section *ms, unsigned long pfn)
ms               1335 include/linux/mmzone.h 	return test_bit(idx, ms->usage->subsection_map);
ms               1338 include/linux/mmzone.h static inline int pfn_section_valid(struct mem_section *ms, unsigned long pfn)
ms               1347 include/linux/mmzone.h 	struct mem_section *ms;
ms               1351 include/linux/mmzone.h 	ms = __nr_to_section(pfn_to_section_nr(pfn));
ms               1352 include/linux/mmzone.h 	if (!valid_section(ms))
ms               1358 include/linux/mmzone.h 	return early_section(ms) || pfn_section_valid(ms, pfn);
ms                 43 include/linux/module_signature.h int mod_check_sig(const struct module_signature *ms, size_t file_len,
ms                301 include/linux/nvme.h 	__le16			ms;
ms                226 include/linux/regulator/consumer.h int regulator_disable_deferred(struct regulator *regulator, int ms);
ms                416 include/linux/regulator/consumer.h 					     int ms)
ms                948 include/net/ip_vs.h 	struct ipvs_master_sync_state *ms;
ms                501 include/sound/wavefront.h     wavefront_multisample ms;
ms                154 include/uapi/drm/via_drm.h 	__u32 ms;
ms                220 include/video/imx-ipu-v3.h int ipu_idmac_wait_busy(struct ipuv3_channel *channel, int ms);
ms                 21 kernel/module_signature.c int mod_check_sig(const struct module_signature *ms, size_t file_len,
ms                 24 kernel/module_signature.c 	if (be32_to_cpu(ms->sig_len) >= file_len - sizeof(*ms))
ms                 27 kernel/module_signature.c 	if (ms->id_type != PKEY_ID_PKCS7) {
ms                 33 kernel/module_signature.c 	if (ms->algo != 0 ||
ms                 34 kernel/module_signature.c 	    ms->hash != 0 ||
ms                 35 kernel/module_signature.c 	    ms->signer_len != 0 ||
ms                 36 kernel/module_signature.c 	    ms->key_id_len != 0 ||
ms                 37 kernel/module_signature.c 	    ms->__pad[0] != 0 ||
ms                 38 kernel/module_signature.c 	    ms->__pad[1] != 0 ||
ms                 39 kernel/module_signature.c 	    ms->__pad[2] != 0) {
ms                 22 kernel/module_signing.c 	struct module_signature ms;
ms                 28 kernel/module_signing.c 	if (modlen <= sizeof(ms))
ms                 31 kernel/module_signing.c 	memcpy(&ms, mod + (modlen - sizeof(ms)), sizeof(ms));
ms                 33 kernel/module_signing.c 	ret = mod_check_sig(&ms, modlen, info->name);
ms                 37 kernel/module_signing.c 	sig_len = be32_to_cpu(ms.sig_len);
ms                 38 kernel/module_signing.c 	modlen -= sig_len + sizeof(ms);
ms                169 mm/memory_hotplug.c 	struct mem_section *ms;
ms                174 mm/memory_hotplug.c 	ms = __nr_to_section(section_nr);
ms                177 mm/memory_hotplug.c 	memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr);
ms                191 mm/memory_hotplug.c 	usage = ms->usage;
ms                204 mm/memory_hotplug.c 	struct mem_section *ms;
ms                209 mm/memory_hotplug.c 	ms = __nr_to_section(section_nr);
ms                211 mm/memory_hotplug.c 	memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr);
ms                215 mm/memory_hotplug.c 	usage = ms->usage;
ms                499 mm/memory_hotplug.c 	struct mem_section *ms = __nr_to_section(pfn_to_section_nr(pfn));
ms                501 mm/memory_hotplug.c 	if (WARN_ON_ONCE(!valid_section(ms)))
ms                504 mm/memory_hotplug.c 	sparse_remove_section(ms, pfn, nr_pages, map_offset, altmap);
ms                275 mm/page_ext.c  	struct mem_section *ms;
ms                278 mm/page_ext.c  	ms = __pfn_to_section(pfn);
ms                279 mm/page_ext.c  	if (!ms || !ms->page_ext)
ms                281 mm/page_ext.c  	base = get_entry(ms->page_ext, pfn);
ms                283 mm/page_ext.c  	ms->page_ext = NULL;
ms                114 mm/sparse.c    unsigned long __section_nr(struct mem_section *ms)
ms                124 mm/sparse.c    		if ((ms >= root) && (ms < (root + SECTIONS_PER_ROOT)))
ms                130 mm/sparse.c    	return (root_nr * SECTIONS_PER_ROOT) + (ms - root);
ms                133 mm/sparse.c    unsigned long __section_nr(struct mem_section *ms)
ms                135 mm/sparse.c    	return (unsigned long)(ms - mem_section[0]);
ms                191 mm/sparse.c    static void section_mark_present(struct mem_section *ms)
ms                193 mm/sparse.c    	unsigned long section_nr = __section_nr(ms);
ms                198 mm/sparse.c    	ms->section_mem_map |= SECTION_MARKED_PRESENT;
ms                240 mm/sparse.c    		struct mem_section *ms;
ms                245 mm/sparse.c    		ms = __nr_to_section(nr);
ms                246 mm/sparse.c    		subsection_mask_set(ms->usage->subsection_map, pfn, pfns);
ms                279 mm/sparse.c    		struct mem_section *ms;
ms                284 mm/sparse.c    		ms = __nr_to_section(section);
ms                285 mm/sparse.c    		if (!ms->section_mem_map) {
ms                286 mm/sparse.c    			ms->section_mem_map = sparse_encode_early_nid(nid) |
ms                288 mm/sparse.c    			section_mark_present(ms);
ms                333 mm/sparse.c    static void __meminit sparse_init_one_section(struct mem_section *ms,
ms                337 mm/sparse.c    	ms->section_mem_map &= ~SECTION_MAP_MASK;
ms                338 mm/sparse.c    	ms->section_mem_map |= sparse_encode_mem_map(mem_map, pnum)
ms                340 mm/sparse.c    	ms->usage = usage;
ms                566 mm/sparse.c    		struct mem_section *ms;
ms                570 mm/sparse.c    		ms = __nr_to_section(pnum);
ms                571 mm/sparse.c    		ms->section_mem_map = 0;
ms                615 mm/sparse.c    		struct mem_section *ms;
ms                621 mm/sparse.c    		ms = __nr_to_section(section_nr);
ms                622 mm/sparse.c    		ms->section_mem_map |= SECTION_IS_ONLINE;
ms                634 mm/sparse.c    		struct mem_section *ms;
ms                643 mm/sparse.c    		ms = __nr_to_section(section_nr);
ms                644 mm/sparse.c    		ms->section_mem_map &= ~SECTION_IS_ONLINE;
ms                742 mm/sparse.c    	struct mem_section *ms = __pfn_to_section(pfn);
ms                743 mm/sparse.c    	bool section_is_early = early_section(ms);
ms                746 mm/sparse.c    	unsigned long *subsection_map = ms->usage
ms                747 mm/sparse.c    		? &ms->usage->subsection_map[0] : NULL;
ms                787 mm/sparse.c    		if (!PageReserved(virt_to_page(ms->usage))) {
ms                788 mm/sparse.c    			kfree(ms->usage);
ms                789 mm/sparse.c    			ms->usage = NULL;
ms                791 mm/sparse.c    		memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr);
ms                797 mm/sparse.c    		ms->section_mem_map &= ~SECTION_HAS_MEM_MAP;
ms                806 mm/sparse.c    		ms->section_mem_map = (unsigned long)NULL;
ms                813 mm/sparse.c    	struct mem_section *ms = __pfn_to_section(pfn);
ms                821 mm/sparse.c    	if (!ms->usage) {
ms                825 mm/sparse.c    		ms->usage = usage;
ms                827 mm/sparse.c    	subsection_map = &ms->usage->subsection_map[0];
ms                839 mm/sparse.c    			ms->usage = NULL;
ms                851 mm/sparse.c    	if (nr_pages < PAGES_PER_SECTION && early_section(ms))
ms                881 mm/sparse.c    	struct mem_section *ms;
ms                899 mm/sparse.c    	ms = __nr_to_section(section_nr);
ms                901 mm/sparse.c    	section_mark_present(ms);
ms                906 mm/sparse.c    	sparse_init_one_section(ms, section_nr, memmap, ms->usage, 0);
ms                938 mm/sparse.c    void sparse_remove_section(struct mem_section *ms, unsigned long pfn,
ms               3574 net/mac80211/mlme.c 		u32 tu, ms;
ms               3576 net/mac80211/mlme.c 		ms = tu * 1024 / 1000;
ms               3579 net/mac80211/mlme.c 			   mgmt->sa, tu, ms);
ms               3580 net/mac80211/mlme.c 		assoc_data->timeout = jiffies + msecs_to_jiffies(ms);
ms               3582 net/mac80211/mlme.c 		if (ms > IEEE80211_ASSOC_TIMEOUT)
ms                 61 net/mac80211/rc80211_minstrel_debugfs.c 	struct minstrel_debugfs_info *ms;
ms                 65 net/mac80211/rc80211_minstrel_debugfs.c 	ms = kmalloc(2048, GFP_KERNEL);
ms                 66 net/mac80211/rc80211_minstrel_debugfs.c 	if (!ms)
ms                 69 net/mac80211/rc80211_minstrel_debugfs.c 	file->private_data = ms;
ms                 70 net/mac80211/rc80211_minstrel_debugfs.c 	p = ms->buf;
ms                112 net/mac80211/rc80211_minstrel_debugfs.c 	ms->len = p - ms->buf;
ms                114 net/mac80211/rc80211_minstrel_debugfs.c 	WARN_ON(ms->len + sizeof(*ms) > 2048);
ms                123 net/mac80211/rc80211_minstrel_debugfs.c 	struct minstrel_debugfs_info *ms;
ms                127 net/mac80211/rc80211_minstrel_debugfs.c 	ms = kmalloc(2048, GFP_KERNEL);
ms                128 net/mac80211/rc80211_minstrel_debugfs.c 	if (!ms)
ms                131 net/mac80211/rc80211_minstrel_debugfs.c 	file->private_data = ms;
ms                132 net/mac80211/rc80211_minstrel_debugfs.c 	p = ms->buf;
ms                167 net/mac80211/rc80211_minstrel_debugfs.c 	ms->len = p - ms->buf;
ms                169 net/mac80211/rc80211_minstrel_debugfs.c 	WARN_ON(ms->len + sizeof(*ms) > 2048);
ms                 18 net/mac80211/rc80211_minstrel_ht_debugfs.c 	struct minstrel_debugfs_info *ms;
ms                 20 net/mac80211/rc80211_minstrel_ht_debugfs.c 	ms = file->private_data;
ms                 21 net/mac80211/rc80211_minstrel_ht_debugfs.c 	return simple_read_from_buffer(buf, len, ppos, ms->buf, ms->len);
ms                125 net/mac80211/rc80211_minstrel_ht_debugfs.c 	struct minstrel_debugfs_info *ms;
ms                137 net/mac80211/rc80211_minstrel_ht_debugfs.c 	ms = kmalloc(32768, GFP_KERNEL);
ms                138 net/mac80211/rc80211_minstrel_ht_debugfs.c 	if (!ms)
ms                141 net/mac80211/rc80211_minstrel_ht_debugfs.c 	file->private_data = ms;
ms                142 net/mac80211/rc80211_minstrel_ht_debugfs.c 	p = ms->buf;
ms                164 net/mac80211/rc80211_minstrel_ht_debugfs.c 	ms->len = p - ms->buf;
ms                165 net/mac80211/rc80211_minstrel_ht_debugfs.c 	WARN_ON(ms->len + sizeof(*ms) > 32768);
ms                275 net/mac80211/rc80211_minstrel_ht_debugfs.c 	struct minstrel_debugfs_info *ms;
ms                287 net/mac80211/rc80211_minstrel_ht_debugfs.c 	ms = kmalloc(32768, GFP_KERNEL);
ms                289 net/mac80211/rc80211_minstrel_ht_debugfs.c 	if (!ms)
ms                292 net/mac80211/rc80211_minstrel_ht_debugfs.c 	file->private_data = ms;
ms                294 net/mac80211/rc80211_minstrel_ht_debugfs.c 	p = ms->buf;
ms                302 net/mac80211/rc80211_minstrel_ht_debugfs.c 	ms->len = p - ms->buf;
ms                303 net/mac80211/rc80211_minstrel_ht_debugfs.c 	WARN_ON(ms->len + sizeof(*ms) > 32768);
ms                307 net/netfilter/ipvs/ip_vs_sync.c sb_dequeue(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms)
ms                312 net/netfilter/ipvs/ip_vs_sync.c 	if (list_empty(&ms->sync_queue)) {
ms                316 net/netfilter/ipvs/ip_vs_sync.c 		sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff,
ms                319 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_queue_len--;
ms                320 net/netfilter/ipvs/ip_vs_sync.c 		if (!ms->sync_queue_len)
ms                321 net/netfilter/ipvs/ip_vs_sync.c 			ms->sync_queue_delay = 0;
ms                366 net/netfilter/ipvs/ip_vs_sync.c 				 struct ipvs_master_sync_state *ms)
ms                368 net/netfilter/ipvs/ip_vs_sync.c 	struct ip_vs_sync_buff *sb = ms->sync_buff;
ms                372 net/netfilter/ipvs/ip_vs_sync.c 	    ms->sync_queue_len < sysctl_sync_qlen_max(ipvs)) {
ms                373 net/netfilter/ipvs/ip_vs_sync.c 		if (!ms->sync_queue_len)
ms                374 net/netfilter/ipvs/ip_vs_sync.c 			schedule_delayed_work(&ms->master_wakeup_work,
ms                376 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_queue_len++;
ms                377 net/netfilter/ipvs/ip_vs_sync.c 		list_add_tail(&sb->list, &ms->sync_queue);
ms                378 net/netfilter/ipvs/ip_vs_sync.c 		if ((++ms->sync_queue_delay) == IPVS_SYNC_WAKEUP_RATE) {
ms                379 net/netfilter/ipvs/ip_vs_sync.c 			int id = (int)(ms - ipvs->ms);
ms                393 net/netfilter/ipvs/ip_vs_sync.c get_curr_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms,
ms                399 net/netfilter/ipvs/ip_vs_sync.c 	sb = ms->sync_buff;
ms                401 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_buff = NULL;
ms                549 net/netfilter/ipvs/ip_vs_sync.c 	struct ipvs_master_sync_state *ms;
ms                569 net/netfilter/ipvs/ip_vs_sync.c 	ms = &ipvs->ms[id];
ms                570 net/netfilter/ipvs/ip_vs_sync.c 	buff = ms->sync_buff;
ms                577 net/netfilter/ipvs/ip_vs_sync.c 			sb_queue_tail(ipvs, ms);
ms                578 net/netfilter/ipvs/ip_vs_sync.c 			ms->sync_buff = NULL;
ms                589 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_buff = buff;
ms                638 net/netfilter/ipvs/ip_vs_sync.c 	struct ipvs_master_sync_state *ms;
ms                672 net/netfilter/ipvs/ip_vs_sync.c 	ms = &ipvs->ms[id];
ms                691 net/netfilter/ipvs/ip_vs_sync.c 	buff = ms->sync_buff;
ms                697 net/netfilter/ipvs/ip_vs_sync.c 			sb_queue_tail(ipvs, ms);
ms                698 net/netfilter/ipvs/ip_vs_sync.c 			ms->sync_buff = NULL;
ms                711 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_buff = buff;
ms               1635 net/netfilter/ipvs/ip_vs_sync.c 	struct ipvs_master_sync_state *ms =
ms               1638 net/netfilter/ipvs/ip_vs_sync.c 	struct netns_ipvs *ipvs = ms->ipvs;
ms               1641 net/netfilter/ipvs/ip_vs_sync.c 	if (ms->sync_queue_len &&
ms               1642 net/netfilter/ipvs/ip_vs_sync.c 	    ms->sync_queue_delay < IPVS_SYNC_WAKEUP_RATE) {
ms               1643 net/netfilter/ipvs/ip_vs_sync.c 		int id = (int)(ms - ipvs->ms);
ms               1645 net/netfilter/ipvs/ip_vs_sync.c 		ms->sync_queue_delay = IPVS_SYNC_WAKEUP_RATE;
ms               1653 net/netfilter/ipvs/ip_vs_sync.c next_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms)
ms               1657 net/netfilter/ipvs/ip_vs_sync.c 	sb = sb_dequeue(ipvs, ms);
ms               1661 net/netfilter/ipvs/ip_vs_sync.c 	return get_curr_sync_buff(ipvs, ms, IPVS_SYNC_FLUSH_TIME);
ms               1668 net/netfilter/ipvs/ip_vs_sync.c 	struct ipvs_master_sync_state *ms = &ipvs->ms[tinfo->id];
ms               1677 net/netfilter/ipvs/ip_vs_sync.c 		sb = next_sync_buff(ipvs, ms);
ms               1703 net/netfilter/ipvs/ip_vs_sync.c 	while ((sb = sb_dequeue(ipvs, ms)))
ms               1708 net/netfilter/ipvs/ip_vs_sync.c 	sb = get_curr_sync_buff(ipvs, ms, 0);
ms               1818 net/netfilter/ipvs/ip_vs_sync.c 		if (ipvs->ms)
ms               1838 net/netfilter/ipvs/ip_vs_sync.c 		struct ipvs_master_sync_state *ms;
ms               1841 net/netfilter/ipvs/ip_vs_sync.c 		ipvs->ms = kcalloc(count, sizeof(ipvs->ms[0]), GFP_KERNEL);
ms               1842 net/netfilter/ipvs/ip_vs_sync.c 		if (!ipvs->ms)
ms               1844 net/netfilter/ipvs/ip_vs_sync.c 		ms = ipvs->ms;
ms               1845 net/netfilter/ipvs/ip_vs_sync.c 		for (id = 0; id < count; id++, ms++) {
ms               1846 net/netfilter/ipvs/ip_vs_sync.c 			INIT_LIST_HEAD(&ms->sync_queue);
ms               1847 net/netfilter/ipvs/ip_vs_sync.c 			ms->sync_queue_len = 0;
ms               1848 net/netfilter/ipvs/ip_vs_sync.c 			ms->sync_queue_delay = 0;
ms               1849 net/netfilter/ipvs/ip_vs_sync.c 			INIT_DELAYED_WORK(&ms->master_wakeup_work,
ms               1851 net/netfilter/ipvs/ip_vs_sync.c 			ms->ipvs = ipvs;
ms               1914 net/netfilter/ipvs/ip_vs_sync.c 		kfree(ipvs->ms);
ms               1915 net/netfilter/ipvs/ip_vs_sync.c 		ipvs->ms = NULL;
ms               1954 net/netfilter/ipvs/ip_vs_sync.c 		if (!ipvs->ms)
ms               1972 net/netfilter/ipvs/ip_vs_sync.c 			struct ipvs_master_sync_state *ms = &ipvs->ms[id];
ms               1978 net/netfilter/ipvs/ip_vs_sync.c 			cancel_delayed_work_sync(&ms->master_wakeup_work);
ms               1983 net/netfilter/ipvs/ip_vs_sync.c 		kfree(ipvs->ms);
ms               1984 net/netfilter/ipvs/ip_vs_sync.c 		ipvs->ms = NULL;
ms               3280 net/netfilter/nf_tables_api.c 	u64 ms = be64_to_cpu(nla_get_be64(nla));
ms               3284 net/netfilter/nf_tables_api.c 	if (ms >= max)
ms               3287 net/netfilter/nf_tables_api.c 	ms *= NSEC_PER_MSEC;
ms               3288 net/netfilter/nf_tables_api.c 	*result = nsecs_to_jiffies64(ms);
ms                262 sound/core/seq/seq_midi.c 	struct seq_midisynth *msynth, *ms;
ms                325 sound/core/seq/seq_midi.c 		ms = &msynth[p];
ms                327 sound/core/seq/seq_midi.c 		if (snd_seq_midisynth_new(ms, card, device, p) < 0)
ms                371 sound/core/seq/seq_midi.c 		pcallbacks.private_data = ms;
ms                382 sound/core/seq/seq_midi.c 		ms->seq_client = client->seq_client;
ms                383 sound/core/seq/seq_midi.c 		ms->seq_port = port->addr.port;
ms               1216 sound/isa/wavefront/wavefront_synth.c 	num_samples = (1<<(header->hdr.ms.NumberOfSamples&7));
ms               1217 sound/isa/wavefront/wavefront_synth.c 	msample_hdr[2] = (unsigned char) header->hdr.ms.NumberOfSamples;
ms               1221 sound/isa/wavefront/wavefront_synth.c 				      header->hdr.ms.NumberOfSamples,
ms               1226 sound/isa/wavefront/wavefront_synth.c 		       i, header->hdr.ms.SampleNumber[i]);
ms               1227 sound/isa/wavefront/wavefront_synth.c 		munge_int32 (header->hdr.ms.SampleNumber[i],
ms               1269 sound/isa/wavefront/wavefront_synth.c 	header->hdr.ms.NumberOfSamples = log_ns[0];
ms               1293 sound/isa/wavefront/wavefront_synth.c 		header->hdr.ms.SampleNumber[i] =
ms               1297 sound/isa/wavefront/wavefront_synth.c 					i, header->hdr.ms.SampleNumber[i]);
ms               2625 sound/soc/codecs/wm8994.c 	int ms = 0;
ms               2650 sound/soc/codecs/wm8994.c 		ms = WM8994_AIF1_MSTR;
ms               2736 sound/soc/codecs/wm8994.c 			    ms);
ms                 14 tools/perf/arch/arm64/annotate/instructions.c 			    struct map_symbol *ms __maybe_unused)
ms                  5 tools/perf/arch/s390/annotate/instructions.c 			    struct map_symbol *ms)
ms                  8 tools/perf/arch/s390/annotate/instructions.c 	struct map *map = ms->map;
ms                 58 tools/perf/arch/s390/annotate/instructions.c 			   struct map_symbol *ms __maybe_unused)
ms                304 tools/perf/builtin-annotate.c 		return symbol__tty_annotate(he->ms.sym, he->ms.map, evsel, &ann->opts);
ms                306 tools/perf/builtin-annotate.c 	return symbol__tty_annotate2(he->ms.sym, he->ms.map, evsel, &ann->opts);
ms                320 tools/perf/builtin-annotate.c 		if (he->ms.sym == NULL || he->ms.map->dso->annotate_warned)
ms                324 tools/perf/builtin-annotate.c 		    (strcmp(he->ms.sym->name, ann->sym_hist_filter) != 0))
ms                327 tools/perf/builtin-annotate.c 		notes = symbol__annotation(he->ms.sym);
ms                621 tools/perf/builtin-diff.c 	if (!he->ms.map || !he->ms.sym)
ms                624 tools/perf/builtin-diff.c 	notes = symbol__annotation(he->ms.sym);
ms                632 tools/perf/builtin-diff.c 	for (unsigned int i = 0; i < symbol__size(he->ms.sym); i++) {
ms                641 tools/perf/builtin-diff.c 			init_block_info(bi, he->ms.sym, &ch[i], i);
ms               1352 tools/perf/builtin-diff.c 	start_line = map__srcline(he->ms.map, bi->sym->start + bi->start,
ms               1353 tools/perf/builtin-diff.c 				  he->ms.sym);
ms               1355 tools/perf/builtin-diff.c 	end_line = map__srcline(he->ms.map, bi->sym->start + bi->end,
ms               1356 tools/perf/builtin-diff.c 				he->ms.sym);
ms                643 tools/perf/builtin-report.c 	struct symbol *sym = he->ms.sym;
ms                648 tools/perf/builtin-report.c 		symbol__annotate2(sym, he->ms.map, evsel,
ms               2762 tools/perf/builtin-sched.c 		if (chain->ms.sym && chain->ms.sym->ignore)
ms                115 tools/perf/builtin-top.c 	if (!he || !he->ms.sym)
ms                120 tools/perf/builtin-top.c 	sym = he->ms.sym;
ms                121 tools/perf/builtin-top.c 	map = he->ms.map;
ms                160 tools/perf/builtin-top.c 	struct symbol *sym = he->ms.sym;
ms                197 tools/perf/builtin-top.c 	struct symbol *sym = he->ms.sym;
ms                202 tools/perf/builtin-top.c 			     top->sym_filter_entry->ms.sym != sym)))
ms                221 tools/perf/builtin-top.c 		if (err == -ERANGE && !he->ms.map->erange_warned)
ms                222 tools/perf/builtin-top.c 			ui__warn_map_erange(he->ms.map, sym, ip);
ms                246 tools/perf/builtin-top.c 	symbol = he->ms.sym;
ms                259 tools/perf/builtin-top.c 	more = symbol__annotate_printf(symbol, he->ms.map, top->sym_evsel, &top->annotation_opts);
ms                412 tools/perf/builtin-top.c 		if (n->ms.sym && !strcmp(buf, n->ms.sym->name)) {
ms                434 tools/perf/builtin-top.c 		struct symbol *sym = top->sym_filter_entry->ms.sym;
ms                184 tools/perf/tests/hists_common.c 				he->ms.map->dso->short_name,
ms                185 tools/perf/tests/hists_common.c 				he->ms.sym->name, he->stat.period);
ms                211 tools/perf/tests/hists_common.c 				he->ms.map->dso->short_name,
ms                212 tools/perf/tests/hists_common.c 				he->ms.sym->name, he->stat.period,
ms                153 tools/perf/tests/hists_cumulate.c #define DSO(he)   (he->ms.map->dso->short_name)
ms                154 tools/perf/tests/hists_cumulate.c #define SYM(he)   (he->ms.sym->name)
ms                158 tools/perf/tests/hists_cumulate.c #define CDSO(cl)  (cl->ms.map->dso->short_name)
ms                159 tools/perf/tests/hists_cumulate.c #define CSYM(cl)  (cl->ms.sym->name)
ms                164 tools/perf/tests/hists_link.c 					he->thread, he->ms.map, he->ms.sym)) {
ms                216 tools/perf/tests/hists_link.c 					 he->thread, he->ms.map, he->ms.sym) &&
ms                219 tools/perf/tests/hists_link.c 					 he->thread, he->ms.map, he->ms.sym)) {
ms                119 tools/perf/tests/hists_output.c #define DSO(he)   (he->ms.map->dso->short_name)
ms                120 tools/perf/tests/hists_output.c #define SYM(he)   (he->ms.sym->name)
ms                 42 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = browser->priv;
ms                 43 tools/perf/ui/browsers/annotate.c 	return symbol__annotation(ms->sym);
ms                153 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = ab->b.priv;
ms                154 tools/perf/ui/browsers/annotate.c 	struct symbol *sym = ms->sym;
ms                305 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = browser->b.priv;
ms                306 tools/perf/ui/browsers/annotate.c 	struct symbol *sym = ms->sym;
ms                413 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = browser->b.priv;
ms                434 tools/perf/ui/browsers/annotate.c 	symbol__tui_annotate(dl->ops.target.sym, ms->map, evsel, hbt, browser->opts);
ms                435 tools/perf/ui/browsers/annotate.c 	sym_title(ms->sym, ms->map, title, sizeof(title), browser->opts->percent_type);
ms                611 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = browser->priv;
ms                612 tools/perf/ui/browsers/annotate.c 	struct symbol *sym = ms->sym;
ms                618 tools/perf/ui/browsers/annotate.c 	sym_title(sym, ms->map, symbol_dso, sizeof(symbol_dso), ab->opts->percent_type);
ms                665 tools/perf/ui/browsers/annotate.c 	struct map_symbol *ms = browser->b.priv;
ms                666 tools/perf/ui/browsers/annotate.c 	struct symbol *sym = ms->sym;
ms                667 tools/perf/ui/browsers/annotate.c 	struct annotation *notes = symbol__annotation(ms->sym);
ms                831 tools/perf/ui/browsers/annotate.c 			map_symbol__annotation_dump(ms, evsel, browser->opts);
ms                873 tools/perf/ui/browsers/annotate.c int map_symbol__tui_annotate(struct map_symbol *ms, struct evsel *evsel,
ms                877 tools/perf/ui/browsers/annotate.c 	return symbol__tui_annotate(ms->sym, ms->map, evsel, hbt, opts);
ms                888 tools/perf/ui/browsers/annotate.c 	return map_symbol__tui_annotate(&he->ms, evsel, hbt, opts);
ms                897 tools/perf/ui/browsers/annotate.c 	struct map_symbol ms = {
ms                908 tools/perf/ui/browsers/annotate.c 			.priv	 = &ms,
ms                396 tools/perf/ui/browsers/hists.c 	struct map_symbol *ms = browser->selection;
ms                397 tools/perf/ui/browsers/hists.c 	struct callchain_list *cl = container_of(ms, struct callchain_list, ms);
ms                400 tools/perf/ui/browsers/hists.c 	if (!he || !ms)
ms                403 tools/perf/ui/browsers/hists.c 	if (ms == &he->ms)
ms                745 tools/perf/ui/browsers/hists.c 	bool show_annotated = browser->show_dso && chain->ms.sym && symbol__annotation(chain->ms.sym)->src;
ms                750 tools/perf/ui/browsers/hists.c 		browser->selection = &chain->ms;
ms               1259 tools/perf/ui/browsers/hists.c 		browser->selection = &entry->ms;
ms               1372 tools/perf/ui/browsers/hists.c 		browser->selection = &entry->ms;
ms               2359 tools/perf/ui/browsers/hists.c 	struct map_symbol 	ms;
ms               2379 tools/perf/ui/browsers/hists.c 	notes = symbol__annotation(act->ms.sym);
ms               2384 tools/perf/ui/browsers/hists.c 	err = map_symbol__tui_annotate(&act->ms, evsel, browser->hbt,
ms               2411 tools/perf/ui/browsers/hists.c 	act->ms.map = map;
ms               2412 tools/perf/ui/browsers/hists.c 	act->ms.sym = sym;
ms               2482 tools/perf/ui/browsers/hists.c 	struct map *map = act->ms.map;
ms               2517 tools/perf/ui/browsers/hists.c 	act->ms.map = map;
ms               2526 tools/perf/ui/browsers/hists.c 	map__browse(act->ms.map);
ms               2540 tools/perf/ui/browsers/hists.c 	act->ms.map = map;
ms               2556 tools/perf/ui/browsers/hists.c 	else if (act->ms.sym)
ms               2557 tools/perf/ui/browsers/hists.c 		len += strlen(act->ms.sym->name);
ms               2566 tools/perf/ui/browsers/hists.c 	} else if (act->ms.sym) {
ms               2568 tools/perf/ui/browsers/hists.c 			  act->ms.sym->name);
ms               2622 tools/perf/ui/browsers/hists.c 	act->ms.sym = sym;
ms               2945 tools/perf/ui/browsers/hists.c 			actions->ms.map = browser->selection->map;
ms               2946 tools/perf/ui/browsers/hists.c 			actions->ms.sym = browser->selection->sym;
ms               2953 tools/perf/ui/browsers/hists.c 			actions->ms.map = map;
ms               2984 tools/perf/ui/browsers/hists.c 				actions->ms.sym = NULL;
ms               3057 tools/perf/ui/browsers/hists.c 			actions->ms.map = map;
ms                245 tools/perf/ui/gtk/annotate.c 	return symbol__gtk_annotate(he->ms.sym, he->ms.map, evsel, hbt);
ms                105 tools/perf/ui/stdio/hist.c 	rem_hits.ms.sym = rem_sq_bracket;
ms                860 tools/perf/ui/stdio/hist.c 		if (h->ms.map == NULL && verbose > 1) {
ms                240 tools/perf/util/annotate.c static int call__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms)
ms                243 tools/perf/util/annotate.c 	struct map *map = ms->map;
ms                329 tools/perf/util/annotate.c static int jump__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms)
ms                331 tools/perf/util/annotate.c 	struct map *map = ms->map;
ms                332 tools/perf/util/annotate.c 	struct symbol *sym = ms->sym;
ms                477 tools/perf/util/annotate.c static int lock__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms)
ms                492 tools/perf/util/annotate.c 	    ops->locked.ins.ops->parse(arch, ops->locked.ops, ms) < 0)
ms                535 tools/perf/util/annotate.c static int mov__parse(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms __maybe_unused)
ms                596 tools/perf/util/annotate.c static int dec__parse(struct arch *arch __maybe_unused, struct ins_operands *ops, struct map_symbol *ms __maybe_unused)
ms               1097 tools/perf/util/annotate.c 	return symbol__inc_addr_samples(he->ms.sym, he->ms.map, evsel, ip, sample);
ms               1100 tools/perf/util/annotate.c static void disasm_line__init_ins(struct disasm_line *dl, struct arch *arch, struct map_symbol *ms)
ms               1107 tools/perf/util/annotate.c 	if (dl->ins.ops->parse && dl->ins.ops->parse(arch, &dl->ops, ms) < 0)
ms               1142 tools/perf/util/annotate.c 	struct map_symbol	 ms;
ms               1225 tools/perf/util/annotate.c 			disasm_line__init_ins(dl, args->arch, &args->ms);
ms               1492 tools/perf/util/annotate.c 	struct map *map = args->ms.map;
ms               1539 tools/perf/util/annotate.c 	args->ms.sym  = sym;
ms               1721 tools/perf/util/annotate.c 	struct map *map = args->ms.map;
ms               1825 tools/perf/util/annotate.c 			args->ms.sym  = sym;
ms               1836 tools/perf/util/annotate.c 		args->ms.sym  = sym;
ms               1863 tools/perf/util/annotate.c 	struct map *map = args->ms.map;
ms               2108 tools/perf/util/annotate.c 	args.ms.map = map;
ms               2109 tools/perf/util/annotate.c 	args.ms.sym = sym;
ms               2431 tools/perf/util/annotate.c int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel,
ms               2440 tools/perf/util/annotate.c 	if (asprintf(&filename, "%s.annotation", ms->sym->name) < 0)
ms               2453 tools/perf/util/annotate.c 		ms->sym->name, ms->map->dso->long_name, ev_name);
ms               2454 tools/perf/util/annotate.c 	symbol__annotate_fprintf2(ms->sym, fp, opts);
ms                 60 tools/perf/util/annotate.h 	int (*parse)(struct arch *arch, struct ins_operands *ops, struct map_symbol *ms);
ms                391 tools/perf/util/annotate.h int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel,
ms                585 tools/perf/util/callchain.c 		call->ms.sym = cursor_node->sym;
ms                586 tools/perf/util/callchain.c 		call->ms.map = map__get(cursor_node->map);
ms                645 tools/perf/util/callchain.c 			map__zput(call->ms.map);
ms                723 tools/perf/util/callchain.c 		if (node->sym && cnode->ms.sym) {
ms                730 tools/perf/util/callchain.c 			if (cnode->ms.sym->inlined || node->sym->inlined) {
ms                731 tools/perf/util/callchain.c 				match = match_chain_strings(cnode->ms.sym->name,
ms                736 tools/perf/util/callchain.c 				match = match_chain_dso_addresses(cnode->ms.map, cnode->ms.sym->start,
ms                745 tools/perf/util/callchain.c 		match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->map, node->ip);
ms               1008 tools/perf/util/callchain.c 					list->ms.map, list->ms.sym,
ms               1011 tools/perf/util/callchain.c 		map__zput(list->ms.map);
ms               1154 tools/perf/util/callchain.c 	if (cl->ms.sym) {
ms               1155 tools/perf/util/callchain.c 		const char *inlined = cl->ms.sym->inlined ? " (inlined)" : "";
ms               1159 tools/perf/util/callchain.c 					    cl->ms.sym->name, cl->srcline,
ms               1163 tools/perf/util/callchain.c 					    cl->ms.sym->name, inlined);
ms               1169 tools/perf/util/callchain.c 			  cl->ms.map ?
ms               1170 tools/perf/util/callchain.c 			  cl->ms.map->dso->short_name :
ms               1462 tools/perf/util/callchain.c 		map__zput(list->ms.map);
ms               1468 tools/perf/util/callchain.c 		map__zput(list->ms.map);
ms               1532 tools/perf/util/callchain.c 			map__get(new->ms.map);
ms               1553 tools/perf/util/callchain.c 		map__zput(chain->ms.map);
ms                119 tools/perf/util/callchain.h 	struct map_symbol	ms;
ms                 88 tools/perf/util/hist.c 	if (h->ms.sym) {
ms                 89 tools/perf/util/hist.c 		symlen = h->ms.sym->namelen + 4;
ms                103 tools/perf/util/hist.c 	if (h->ms.map) {
ms                104 tools/perf/util/hist.c 		len = dso__name_len(h->ms.map->dso);
ms                428 tools/perf/util/hist.c 	map__get(he->ms.map);
ms                501 tools/perf/util/hist.c 	map__zput(he->ms.map);
ms                613 tools/perf/util/hist.c 			if (he->ms.map != entry->ms.map) {
ms                614 tools/perf/util/hist.c 				map__put(he->ms.map);
ms                615 tools/perf/util/hist.c 				he->ms.map = map__get(entry->ms.map);
ms                691 tools/perf/util/hist.c 		.ms = {
ms               1064 tools/perf/util/hist.c 		.ms = {
ms               1244 tools/perf/util/hist.c 	map__zput(he->ms.map);
ms               2044 tools/perf/util/hist.c 	    (he->ms.map == NULL || he->ms.map->dso != hists->dso_filter)) {
ms               2068 tools/perf/util/hist.c 	    (!he->ms.sym || strstr(he->ms.sym->name,
ms                456 tools/perf/util/hist.h int map_symbol__tui_annotate(struct map_symbol *ms, struct evsel *evsel,
ms                489 tools/perf/util/hist.h static inline int map_symbol__tui_annotate(struct map_symbol *ms __maybe_unused,
ms                191 tools/perf/util/sort.c 	return _sort__dso_cmp(right->ms.map, left->ms.map);
ms                209 tools/perf/util/sort.c 	return _hist_entry__dso_snprintf(he->ms.map, bf, size, width);
ms                219 tools/perf/util/sort.c 	return dso && (!he->ms.map || he->ms.map->dso != dso);
ms                265 tools/perf/util/sort.c 	if (!left->ms.sym && !right->ms.sym)
ms                278 tools/perf/util/sort.c 	return _sort__sym_cmp(left->ms.sym, right->ms.sym);
ms                284 tools/perf/util/sort.c 	if (!left->ms.sym || !right->ms.sym)
ms                285 tools/perf/util/sort.c 		return cmp_null(left->ms.sym, right->ms.sym);
ms                287 tools/perf/util/sort.c 	return strcmp(right->ms.sym->name, left->ms.sym->name);
ms                328 tools/perf/util/sort.c 	return _hist_entry__sym_snprintf(he->ms.map, he->ms.sym, he->ip,
ms                339 tools/perf/util/sort.c 	return sym && (!he->ms.sym || !strstr(he->ms.sym->name, sym));
ms                355 tools/perf/util/sort.c 	return map__srcline(he->ms.map, he->ip, he->ms.sym);
ms                448 tools/perf/util/sort.c 	struct symbol *sym = he->ms.sym;
ms                502 tools/perf/util/sort.c 	struct map *map = e->ms.map;
ms                508 tools/perf/util/sort.c 			 e->ms.sym, false, true, true, e->ip);
ms               1587 tools/perf/util/sort.c 	return _sort__sym_size_cmp(right->ms.sym, left->ms.sym);
ms               1602 tools/perf/util/sort.c 	return _hist_entry__sym_size_snprintf(he->ms.sym, bf, size, width);
ms               1626 tools/perf/util/sort.c 	return _sort__dso_size_cmp(right->ms.map, left->ms.map);
ms               1642 tools/perf/util/sort.c 	return _hist_entry__dso_size_snprintf(he->ms.map, bf, size, width);
ms                 96 tools/perf/util/sort.h 	struct map_symbol	ms;